var/home/core/zuul-output/0000755000175000017500000000000015135627030014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015135634056015501 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000270267115135633772020301 0ustar corecore7wikubelet.log_o[;r)Br'o b-n(!9t%Cs7}g/غIs,r.k9GfD 2|I_翪|mvşo#oVݏKf+ovpZjtC4%_̿f\ϘקjzuQ6/㴻|]=ry+/vWŊ7 .=*EbqZnx.h{nۯSa ׋D*%(Ϗ_϶ݬvGR)$DD D~m{]iX\|U. $ॄKЗ/83Jp ώI8&xėv=E|;FmZl8T*v (6pk**+ Le*gUWi [ӊg*XCF*A(-aD~JwFPO7M$n6iXύO^%26lDt#3{f!f6;WR.!$5 J:1*S%V!F([EbD]娍ԹiE03`Cfw&:ɴ@=yN{f}\{+>2^G) u.`l(Sm&F4a0>eBmFR5]!PI6f٘"y/(":[#;`1}+7 s'ϨF&%8'# $9b"r>B)GF%\bi/ Ff/Bp 4YH~BŊ6EZ|^߸3%L[EC 7gg/碓@e=Vn)h\\lwCzDiQJxTsL] ,=M`nͷ~Vܯ5n|X&pNz7l9HGAr Mme)M,O!Xa~YB ɻ!@J$ty#&i 5ܘ=ЂK]IIɻ]rwbXh)g''H_`!GKF5/O]Zڢ>:O񨡺ePӋ&56zGnL!?lJJYq=Wo/"IyQ4\:y|6h6dQX0>HTG5QOuxMe 1׶/5άRIo>a~W;D=;y|AAY'"葋_d$Ə{(he NSfX1982TH#D֪v3l"<, { Tms'oI&'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oTW⇊AqO:rƭĘ DuZ^ To3dEN/} fI+?|Uz5SUZa{P,97óI,Q{eNFV+(hʺb ״ʻʞX6ýcsT z`q 0C?41- _n^ylSO2|'W'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8c3ilLJ!Ip,2(( *%KGj   %*e5-wFp"a~fzqu6tY,d,`!qIv꜒"T[1!I!NwL}\|}.b3oXR\(L _nJB/_xY.# ſԸv}9U}'/o uSH<:˷tGLS0l/LKcQ.os2% t)Eh~2p cL1%'4-1a_`[Zz㧦|k˭c ĚOρ_} Ewt3th?tvͪ{~;J0= |JUԍ;Iw}/9nh7l%>'ct Հ}a>-:(QxPyA Z UcÖgڌ:8cΗ|U1,-N9 dI [@3YN%:ò6PT:”QVay 77ĐrX(K&Y5+$wL#ɽ 4d-bbdAJ?w:P>n^2] e}gjFX@&avF묇cTy^}m .Ŏ7Uֻ󂊹P-\!3^.Y9[XԦo Έ')Ji.VՕH4~)(kKC&;嶑, }t&&\5u17\I@ 5O? ʴ(aPqPϟ'+'h=TԫeVN^_'ƟlQ.Ff!bpRw@\6"yr+i37Z_j*YLfnYJ~Z~okJX /6h\%WyΈ W\Of[=߰+ednU$YD',jߎW&7DXǜߍG`DbE#0Y4&|޻xѷ\;_Z^sнM\&+1gWo'Y;l>V ̍"ޛ4tO,{=hFѓ$b =D(zn;Y<1x~SJ^{vn 9 j1шk'L"cE=K]A(oQ۲6+ktwLzG,87^ 9H\yqū1)\(v8pHA"ΈGVp"c ?Z)hm.2;sl$瓴ӘIe~H|.Y#C^SJĽHǀeTwvy"v܅ ]?22R.lQPa ˆSܫ1z.x62%z].`Gn&*7b/!~x]y7D7@u邗`unn_ư-a9t_/.9tTo]r8-X{TMYtt =0AMUk}G9^UA,;Tt,"Dxl DfA\w; &`Ͱ٢x'H/jh7hM=~ ֟y[dI~fHIqC۶1Ik\)3 5Ķ']?SؠC"j_6Ÿ9؎]TTjm\D^x6ANbC ]tVUKe$,\ܺI `Qز@UӬ@B {~6caR!=A>\+܁<lW Gϸ}^w'̅dk  C 7fbU{3Se[s %'!?xL 2ڲ]>i+m^CM&WTj7ȗE!NC6P}H`k(FUM gul)b ;2n6'k}ˍ[`-fYX_pL +1wu(#'3"fxsuҮױdy.0]?ݽb+ uV4}rdM$ѢIA$;~Lvigu+]NC5ÿ nNჶT@~ܥ 7-mU,\rXmQALglNʆ P7k%v>"WCyVtnV K`pC?fE?~fjBwU&'ᚡilRї`m] leu]+?T4v\% ;qF0qV(]pP4W =d#t ru\M{Nj.~27)p|Vn60֭l$4԰vg`i{ 6uwŇctyX{>GXg&[ņzP8_ "J~7+0_t[%XU͍ &dtO:odtRWon%*44JٵK+Woc.F3 %N%FF"HH"\$ۤ_5UWd̡bh塘ZRI&{3TUFp/:4TƳ5[۲yzz+ 4D.Ճ`!TnPFp':.4dMFN=/5ܙz,4kA<:z7y0^} "NqK$2$ Ri ?2,ᙌEK@-V3ʱd:/4Kwm2$'dW<qIE2Ľ)5kJҼMЌ DR3csf6rRSr[I߽ogCc;S5ׂdKZ=M3դ#F;SYƘK`K<<ƛ G׌MU.APf\M*t*vw]xo{:l[n=`smFQµtxx7/W%g!&^=SzDNew(æ*m3D Bo.hI"!A6:uQզ}@j=Mo<}nYUw1Xw:]e/sm lˣaVۤkĨdԖ)RtS2 "E I"{;ōCb{yex&Td >@).p$`XKxnX~E膂Og\IGֻq<-uˮ◶>waPcPw3``m- } vS¢=j=1 W=&;JW(7b ?Q.|K,ϩ3g)D͵Q5PBj(h<[rqTɈjM-y͢FY~p_~O5-֠kDNTͷItI1mk"@$AǏ}%S5<`d+0o,AրcbvJ2O`gA2Ȏp@N ̢t-mfeF;gUаQ/ .D%ES*;OLRX[vDb:7a}YF30H #iSpʳ]'_'ĕ -׉6tfЮ$zͪO_sYq+q艻*vzh5~Yy;,DiYTP;o./~^.6+zZFD& m@WXe{sa 2tc^XS?irG#^ŲDI'H_Ȯ;RJ&GT.Kwj;of¬zHmmS2ҒN'=zAΈ\b*K ڤUy""&D@iS=3&N+ǵtX^7ǩX"CA⥎å+4@{D/-:u5I꾧fY iʱ= %lHsd6+H~ Δ,&颒$tSL{yєYa$ H>t~q؈xRmkscXQG~gD20zQ*%iQI$!h/Vo^:y1(t˥C"*FFDEMAƚh $ /ɓzwG1Ƙl"oN:*xmS}V<"dH,^)?CpҒ7UΊ,*n.֙J߾?Ϲhӷƀc"@9Fў-Zm1_tH[A$lVE%BDI yȒv $FO[axr Y#%b Hw)j4&hCU_8xS] _N_Z6KhwefӞ@蹃DROo X"%q7<# '9l%w:9^1ee-EKQ'<1=iUNiAp(-I*#iq&CpB.$lٴާt!jU_L~Tb_,֪r>8P_䅱lw1ù=LAЦz38ckʖYz ~kQRL Q rGQ/ȆMC)vg1Xa!&'0Dp\~^=7jv "8O AfI; P|ޓܜ 8qܦzl5tw@,Mڴg$%82h7էoaz32h>`XT>%)pQ}Tgĸ6Coɲ=8f`KݜȆqDDbZ:B#O^?tNGw\Q.pPO @:Cg9dTcxRk&%])ў}VLN]Nbjgg`d]LGϸ.yҵUCL(us6*>B 2K^ sBciۨvtl:J;quӋkKϮ듃ԁ6Y.0O۾'8V%1M@)uIw].5km~Ҷ綝R(mtV3rșjmjJItHڒz>6nOj5~IJ|~!yKڮ2 h 3x}~ے4WYr9Ts] AA$ұ}21;qbUwRK #}u'tLi'^Y&,mCM)eu㠥Ѻ\a}1:V1zMzT}R,IA e<%!vĉq|?mtB|A ?dXuWLGml?*uTC̶V`FVY>ECmDnG+UaKtȃbeb筃kݴO~f^⊈ 8MK?:mM;ߵoz+O~e3݌ƺ(ܸf)*gCQE*pp^~x܃`U'A~E90t~8-2S󹞙nk56s&"mgVKA: X>7QQ-CDC'| #]Y1E-$nP4N0#C'dvܸȯ.vIH"ŐR ;@~y>Kv{) 9AG ćͩ$.!б~N8i"1KФ\L7/,U@.ڮO?mُa ې!rGHw@56DǑq LA!&mYJ*ixz2*{_;IYJXFfQ* 0kA".mݡ"3`Rd1_u6d逖`7xGMf}k/⨼0Κ_pLq7k!dT x삖A7 u/~&ӄMu.<|yi I?@)XJ7{ޱ?Q]{#\4ZfR-dVaz./f+yGNMGOK?2_~3\z=y}^G$*A! IcuR.o=MZ9zu b#s9@*иrI@*qQN||Ix;I}&ݢ6ɢ}{]x}_o>Mm8S]~(wml"Ms>\΋"?|NKfֱn !l+ a[s׭dֲcUh=Ɩ9b&2} -/f;M.~dhÓ5¨LIa6PnzɗBQiG'CXt!*<0U-(qc;}*CiKe@p&Em&x!i6ٱ˭K& FCfJ9%ٕQ·BD-]R1#]TROr}S [;Zcq6xMY 6seAU9c>Xf~TTX)QӅtӚe~=WtX-sJb?U'3X7J4l+Cj%LPFxŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9U_R=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8|Ldl?6o AMҪ1EzyNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*6mMq!R{QWR=oVbmyanUn.Uqsy.?W8 r[zW*8nؿ[;vmcoW]"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍeys=.EinE% 1zY\+͕߬VͭW_겼cazyU1wOw)Ǽn@6 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U6IXabPde Ӽ8Ģ8GɄnb'G ֤Mcv4?>HC78NE@UMc8>`TvZwm_1~lQ^4is].%RnNJIx5mf8 ək$1ZɅLU$i,*γ'[Jd f $MbJ/4_3?V5(x*.<"*D)E\,2SicS-1VEd"HZ'0ZŨy4eUoΛ;]Yǹ,TGSmX&Bf~j2O`R<**2jPk;V3X:,o$M3M-"uo&&Nsl'<WW5z 9> K2>g"%3SdZT `Z8 gB^Oz;GI~5V s.)]RbtOlɮ4#mEL//;EBe224CjKeqe7t_WEX$]~獀eqEm"YD*4#ўu"C|p0\8q#"ӚmY,q^{PE^3t1s^E'_AW%B|9Z bD$`D40O\Da~hǓi^d2xAǮnP%^H} =ßk|A:y? ; {I.R扚wߕ%C|JbB%C|SS-A |ex7 noQ:#yD(-aCo%&PȐ;*cp>N-Ӳ/LbsaԞRx cͣ: zԂRG,(oDb$oWr,:.u нzd$x/}d0!>1>CM˵])DSkT-?8V[lq(lxICaw^LfU}GD2Gƚ5SiLy Rf! `RgW<*'"o(.c>r'&U!z_:tr.dQEzvOwEVWլ%Ku%Q{]jW޼:Dy9Ia,&qJ-pUel(ZinHC4e}KJ'auaTGD\L'IX4_@ $j/K x;K?G0:+b{j@!8)bz|. [F^WT<'&xwK 'ݕ<m)%hL- V[Aď 2hDFy%bTJj]t*Tïn9؞?q癮魹$Wx1W$\tRL9FyV' l;9%/?sU?ǎMv@{0xgwɅbsloc> ZPY*DO'<^tqr-/^8ߠ00w| v?rAr.AkPk7yƋj((ůgծ>u1ۖru%G=iEݧC{*eG1>cXRvs >vb D6cnB3:=_o0`%zn?~ aAlxS#!aT[w[*[j[ȶA]cT=TUj R)olznaq1 C,Xo 6 7<᪆ ϰFeBL n!H61H;R< &R־ndSe zu~)3J<<#e( b6ˣSl9jj\*:hj 6Ծ;@b6Թ  &m9G"p6 < %s-)GLԞ%Uo[VW0^'aE O.LY-ŪQ5qn;e# +b9ԅȊ\Հ2. Rmy'fe69!=̵2s3WaӒ8VK3qts%]\J$`LYw͍Uo#3Mﱵn[D1M膭G]D1aRfJ{߻cd?'iK]9d/ڡQ^zRÔ"o]#zi$@}٧'lOJ4S߲;WPQi|f@Է^uĠ>f).^M:@t{ԝmd&03m4H=iLAuf2-si=(LU' {*wmՙ:OUwu()0PKC1B螴vfLP{Ә)ADvvcW) #rnR'>f g}· 2D6FaZNRu`4Bbb5GA2":C@8@=ٝߩ}7k n" L7;DY%ڍ \YmL[UF[,`7$5rέ/{Wܶf^>~sk'qg4EB6I#%;>OHd0D[{njY$v,/±8Z`{#fd,;YL˼25q2 ai.ӂa^5m'ۄG_|I8ýz{\KUghѯ $4- Q'먶z y@&쏍6W<+ti (yy|1d4  -0EG]8+B)&=x?5q`W-؊20JJ1H⡏Os;[S\kų9ޗo8k>~ 7Ž'$[qC.?yQop}Ji}##pX>Nۤp/+-뛬ݪw6s . w~>:8>u_d`,f4pRhzPb3"(w'sj뻻~GQgjX%" et 5Tٚ"m/c7(q_äѦfKfEn'jP4<ٚ)qZSbՊP얽mx-A1 J OQo6(6忥C^[/q5BY^>M Fk7:E,~j`4^K!PwK)[oSh[1:=jtzThW6Sc"$EB@(DE@ΫZ"EVqيxv\iPh~?O{op5򿼌"^-_[5M~PţD+k/} WRbG^V\Am+b SCB;.NJ8#ՠ/y@ݦ播|PjEPNKy`X퇲)ZJk3塒ͅ_^_\o5N5Z0?_cLJоoVnMK/wWt]_0,tDc;t"^3-82bm W%,sa]!]@~ƘVvvB*N\'QiPJ=Ig(>ƾ$,{߄{n]6 0ܟ~R*q5 onï96 @q^ܩdŶu+{i9PYC} jh@O,q$I:[5ip6r;cD6~وfe+!(H$u09Efv4ryfZZ0bZW"PNq=0>1,L22 nl0N"`%)EHڟIׯzw0:3Sɩ\Wwy5.yb@򼟤pH%G ҷ%L>EykuڱO+xYUa3,B6ꑿS!Kbc.˰@)Jpy 炘bF5GyE,כ(DypU1 ́D9:$?H9:rx(| .gj؏x|8m;p9 MLdY `yt:-,P%f&/쏐Zj~#Ek Ri VDx^+ꊚAׅ<Թgy0^}eӁcY&A zneYQ}2f}TT nF9;d^w\L}h4WsU2"vC}1̫3 @B!qt~Bn~ѠcԣM6I9۬/>f]PgXB9aD<#j t>Ot!rDz2Q=<9%ut:a/2g9sЋ 1ơU-Qgo*3Jܟ\oe`sL:xȆЄ p=(avq>9 ҭg8[pK7a,?,dLJ`>r2p7RZ*E5YYÿέXՀVUu p116bl@XjyQVժ£ LR)twkk@"vYC }9w7--29;Fm-rtvivr.@~Omg79,n^KlV,wS?94GS^4AgepO\ݠœ? >Lvz2'aqoYz5(Ы,a<F-vA.p~7IG33}] ݈GBq^q~v f`6,7\n=[ =Ďm{qM 5D~?-Fg,XA`- ړ)BKWp7)ykٌnxkx7Fՠʥz[@5a$3:Ec'[C!4I 0qP#Jltmj58\E54eVmL,gzg14"V#QtV3! |x4Kɒ-.b3˵<ӱr?M{dR ߱d_yPj1:c:Ka+q$Vuq'\F>7n2#J"]*;~6>.eBnQ7=j^N".34hPv]4 PPjLR\/8 a+B*9y2EBYBu)kҧLS3{v+kԯ;QϪtsVsJрX9H5)5jj6 lNgV,۴oؿ_=ʱEs&S}ϡ{ 0˒wtUle>\y}߂ͪaSQQr 0ȯg'ՕPrt9l^@~ήd, x)P:d0hAЖi.8 21~eI:Ϸ0L]ߖ]PZoa(E:hqr|'xDWi`eh_a.Y[ ߆Ez$_)t'dC?[nkM}~'NS @qAϥ'[ 3gC>^l<=) @ EGr4¥ X5'x뱻<r̓h +8`VH>fۏ%YYt@XV"Eu*fz|ZQQ$$):`8`5Y K pV煵K 9[mDٓ(+>:JaezX8ub$tPtйeC>YH`\8qE1i."iO.Y"7k)kPeWPB׌}¢%kCY1I~?)g!?P0å\) Txs" fܾ@ #_M72ŰLJ_dL^5ϼr1 A}"Br0DɚPZ3) tۏ|0QÖd$]Aa= q |Ny$< 8VO}c`BH8=PҊ*ąN̎Y(Bء[~YEjkOVO z,ww@e|!rv{7i> [ tk!(vx#(ɣ<уcp.bCep"=,۞3-H&aͲMY<փdH3 Ru~M5fYϡoyqtq"lt=<{PдAaGҪ1?21ɛ@0Ǭǫ.AJ'`؀sਧ;9ڮ'YeYq$>E\Y6KY(.\PMޔYgu{"HDV:i5X?nI^UtarBZaRئd t,o9ty^mD#JU]ZF-_8ȃtCcرe f[><6 a̳{h1|8*DӉ* hH/&]TqAy +p`h #o2hYA ?d_o(uOrX6ȏs>Fd(U@.gK{*Ղ[ܞ^t0$bZk^]6 QK "~8J'3OAT ߠ[7qoӖgQ.A%G:p{[pt`^,y==9bq휣3{rj5ÉZBX,m%nN=b~U+8rSi]hk3fd1f`4icՇ]^*+f>wWlÒ8[ѧCitVs܄ůb-,jnm:Kc1YjzPmx5r]h3"JJػEhZ履 >uH/5K6vU9Gcda~{f=?Q7ua:,Rվ`]XP4 FL^[E썞hv ~KM-~u`=M Ap| r sZ]od=1ɻƎWU2>:ێ.AY @!zԩ4.2EjVaQfJ 2##:y668-g}⢈D#64ߛ?S z}E< `{ _) ی#?xÓ|&b=v<,7[NMh#l{}nE a`]mDޘulA6)>`7[ =7ͳi/?>pl:`4o=lE^IoM(nYFo#]-7?Ntof&|DqCg/Hn` C?zK*Jf&zY %h*%ͤOJ%mIlnnC->e_%@^DAxXi3 AOĢL. Kl饸[8gxw?d$)Z-Tc}RMdemDP:w1V%([;8+U36 |۷f#?#G a(- s 2gQXTSra0;7^ ɣ*s(Y}1.S(rr͂~}SB`P [ɖQhNlG/0ۇ].'3`z^C-2rb>jU 20P@犱BFhȥYHIuAmhbLt"ƽE:@^ i8R(|\gdܝ 95Z*Mk_\ҙ#gry4:΍]g5 !?2VX'4kBw;N, s~p tT) -fJHK &H@xLqLMs1lVab9N,#1:GRQwbiǯpdjPB&9C/,[MU@7\.'nE:a#b; }ɥ]{5 Eqb‰3\!t;gW~c$ޥBжY)ٔ[EwK>Ub6Y|{W駫}Nt|trN^oY?Ƌ[&NB9ԓ(Y=,Q?"@p~lle/ Ȃd\[ 1^gX ?in^<u Lj *~MwR%.uP:sl= q A2z٣K䨱xGxiӮX s$8v;le͕ Cmi6|hNoNƘ WNBVw[UP(Ƚ=sK7lLg՟ܥ#QAc %!CnM:/ZԶӼ? %̸\]v,l|ϝK.i]r'5i}s\<5\QNK35V%:YmvciɌxc\V@o]MdJt.{"2fW4B Vۑ=x'n}+&ܩ#upɠ.)0NAh+P2vJ3,1{nh^"'3T 37/Vkxz1wOK+9yC3tPiQ!b%m -*,uNdQ՗5AM^c P@Ice0:xP{#3kNp=\߰h.ky/ޤae@?s_5t-|CJ2Ƕ \FIQGQZF2g+5%K\$5 %F/v7\lrP= 6ZI܀Mnyu6cV+ 5"iG~/,u3\qasKը\r"^TaQd>JWղ.faO(<6&e?1[?JM['ޗ|,a,Nn>y-ֵ}1lC/$߃V0PP0 rr[zu;vZ7)fDׂoN4_uް8&Yn8lFnkF݅,svuǂիE֖R:9I_YBi=y, *jӲLr.Asp 6hD$x%ODjhE> -i54䩕R0h &bH(KS}J{%8=|ދegh_F7!QWUBU&h+d&e" eGWǂm!o뭉feclZ9wUoXF0: 42,2ZWr:}J0̂c=A] 4EgZ ۛ Ng^:擔w,C7y/g8.ɝTC3qC<7m>8{Y'lc+c,'Yjnr=܊!{Np}nru-KKMrU9IWB}:{Ӳ!E7,{QZB@?1 "O&mg]֭S}l,~rN܋UCAU=ܲ޻!դV2iNO54KN0c!ԙ3(оiW+|p@1 /JҵzzZowl4&קf/ݭQ{ .N.O[8`Ru`',beu"}jR'IW';7pd@RQ6#5yN}v`/3dQH4ͽl8ڼ,\c8Ps Dp3~r˂Kuټ(&:wf~PfOQz.<8N>;a{PkyG5T{ԁ{Za(';zςkgv iÎ8J~T9JTԩ'[[),8N%'R\@3\]i]oZ i!^˂\6)- }Di@>gQXm!P55:vd@G|bN?|L@~)ssm@D,&W@}ƀG3-ĈTQ-Iz)qBi^ݨ&Da.5e S~9&PYDcrZꋪ/ )fi@ԓZl Ll c7Zg&.$Ph3 EN1r1@4z(1%ɋjCZ@)H6C_󃑲^3Ùj !}6{Q{4&|wrW'ɟ;]l!L2sseI|Bʽ?֠tQg9*Zb B\?e)U}۱g_+nlN>[zF]Tk"JNs=&/h|E[D. )نrS;p?u̓T}:& [e;;3?Ӳ] 2wu6F 4TT7 ֈ#7v)ag] {8𖸆t <*-vJ/1)R 1,} v]K0l:(df<,̇ Џk;~FL)2I&S<︴Oyy-|x71-nfvߗoyB3qk;7MprU"d?&!_]:@%{=ЬڠJun:s <5FE9C5W7>DRF?|qrSsu-42Ɉw ӫi$x Lf*4xa&uDuWm3Kk&,b똗 bOWt0-@>'n.~ xΛˏD*bWf՜YE\ ]˺̼5NӋO:3(.WN8ce^6V}4^nϦwV=m5H8VS E> k y^;վ 9]O"Nխa˿oM*uL(~}.177fSXiW3)l zdE'ȗ?OR>󡞓)>g &is_7NfvGָtp_>X9U:MWc7#*6٢q3Xcޚq$.m7E F˨if $Q9]%n)uЌ foCEx)bO5O 5Q9UV_}~v b8Ul&s"d؂Z" vYMPAA6A FoE DVH%%-07PmBY>t"eACk?ۍ^g}y+ەz ,iMEZ%n$H[$mra?8iHV[@Zl^@kZDZKylNbU߀bnne 5V4&S,SeJYmJܠ|k%}}H E͞?Ƒ>U &sF}e-G;0nPDYRt77eF´@7I՘Qo_Cۃ.*9 ,cJ ;ɜr>-|ɾkR+$U%L,G'V;) i 1oKv}ߙdԛƺgŸ^Djx#=c0^:v@/=hѦ RIGfTXT`#)EԞ9Cyދ;@xyw}`VfV#rbrKI.h9J<*9Դ:(,]O%}4@yuRsc{#' {)d g@uWOJJ{ZqO:>AN#cŏ3s_ - cZXi`za *d,=R9GMy]rSKw!q$t. ]=TB2xW_f{hRϤ'w=XZ8 u1EVI`ޞ,I\21]?*hi0պ*ç8|mѺi#bTpfҿHp&4'5{Kza l~]KWTeJ$T '_jeQ`5w]}J ZAӞw5 O!FuX ttz7x({%ph/R~[1x7ԮK0Ms,L@b'}wl/H-4xջ Zk5 vq=<6Rv Gy0XTP[X\ghORwĮ6^{u<1-pMЏ?}5 *}Ġ W< p ďb[Dz*x~9a-嘧T*ZS٨E>}=, }}iIRpY]C*ÄlYV1{Hq`kbTϓ~ ^LQlaEdT=ǥ+cn}V:ZhS,cfȯɇ,\O k,tw^z&Epb6&]F!Ҩ  g &hmX.sqClXF0NJL 0*X'!崋h~wh`|r= .Ő(A°RoZ9l3!Mm4 o.zi84Obڒw4>W廒&Œ |C48!ߗR)V<$z4xc2ȭɓP@5tA+{#Mi͏wpK>U0}sqvIg~s܊rsܻw88.-!Xo' \_5x}ر=}q}zaj=ytW#V tޭ|w6bL}Hv>HKEMź%wᤗ!!Z8 !(Ȑc\`@th)Aޑdᔷ8HokY#W#>ZV@YfP1ZUn[X};w+L+v-cNbu1˩.b!:FbNi w2+2֎4F V.\l]ԷB6m j l3+PӊfhG+FLn38o~wD"BLe8Mɺ%g!I;0es s6AE eC&Dƙqzo1u~1MO]BK7 (?zBFȅDž$q{ڰDeZՅzb |acԆ1%~p]=u1K \ܵA޷ecYTZw$6>M_ 2 ܰ+|C"N0m̗WSQUD)SJK7oWA[-q`c5 9A!dZϹvh i=aNFl-/^ `8HSLQ/F>V> ETSm}34kfL)!A*jY74N(£y8 [4U/ t[.2cpAEd@DS^:{-|?N4繊NF` (Ded mBHELz.`_|.W 1bD3S˚:`ݼ~.!m0bp)$=z6 r6| i ? I\ P 6H(MDKy* JgACU:?Q1%mN@\ZCI)uPb+ІEn#^#i+)f I _I7)ʞ i~_!,*g#eZGv az$-eR\gNHXHHhC$.2%;`U7)%%SjX٬ꮪUuwUGPp,1!0Qn$J-IDc"sw1h(4w`NG8wyJNw9#kVj: ͓}֛({^ֳJ'pTK鮽j&.n GH-Z1}U9eMFCj>&ƍ^7t({\L`"a2 Bb8XL 7Z + H-ȀwuU9٥t`v4m6كq^`kG̓7K5x 06#Hwhf$BYYH#PtDP b@U**^%C!xñS;] bc8(ehbQcdJkDDCvV*fXsPsTb8MLA?H&KM")Kt'fPRumydjZbR`P–{=rnנV!dA!z8\ %%'}<fTNx^o1C-DF c:*R*b$I "ND1\! #Qsua;*d\Ld\LN [`-#Հ:ԂK:% T.I} =40 VGUu&3͏EPOHuߖ#h{SeMr5 iSeuVRc7R8e߇?M"Tgh&dKZ~D2)jh,r:z*% .I&+Ҡ@x u`}LJj+D-2dCVQgڶ" 'Z7kl/Gr'" ~rz˶m{/}ujJ^+[sg!}ʋ kK@~lKoYt+?H?WB j\7E2\Zә?Ip>uƽa-fazʢ6moO{y/97(6dؤCqv0H'dtam\znGf*se\=g9X w1-uF\s7^,L=InrOR Oa3=r%*A4IwXA[dzk ȳ6=RV{u%gOηrH%H!cY#T9Q l%s65-6+X>6\ wѧ\xu_Vq#rqrOϝ_; 01`*Cd$5#Bu3/dzSRՍa[ʱ'lb#ɐѾcibHE[cua;a>̦R1-VRrnplx:/8,)ng-*9R6+LM20i0= ݻM3~xžTA`6YvLݧˆ37J- +mjnN]lNmr5j1TZ:**RoӞ吲BV>bdToJ$W%SSLn"b˦kjXe>G-*A封֬K<6AfrͲh<h3!}V&'^;peb%FnnV:pzyx:?>[-LV3,S0fнB0<\@>W譟,([ l[7y?}kJEx41vCp֗ٓauP~\ٱth[+?`ևF Xӝ/_h4huPW/#iގwjKK+Ef/h*tK'Mh T* *n$\Ҿv^}> cJYp))`T!*Soi(ak@йHsDls3c%bn ֣ХD 0sBv!I.E>FK!ZƵWZkc\֫hITڭPCY-z4HRJ)4MQ#S_#MM0ȼIRݩ̛_Ŗ3"Jh9OnH=YfqyT zp0M(TmuqjXx8w*M$}º p+cH)V6:m[6xb/ }3seZ3XY)$Fc5?x]6ɞ}(dR3L(NS}}U\(16'Z/f5>꼭l4l_On}Ljf?p'g6/\,eW蓱Y< 绘}Cƒq/OMƜgFzeAovгKc{mrOE½K͍7I͝ӄ!Q*|\ wu^K پf%7Qs{n,<+9"[s©墤{J7/ٜCϒ`٭vΰ捓דeY|k+i@^=0"&؅dמª(~ǃ n^M'נSXr1_ǒ( 3b_KH)p x#X>ag1c6jTL|:-')``,ByA'9 #FBiTY]+{\YNSZ"ikF\jMTU(X-ՎI R,Ai$qD Ĕ lN-=h)[oP#~e:uJKs]6_2ڟ:kYiy;cP}: 6T=#WXogކFYމnv10FôQ[s 8m^toѱ=3+[el18@t-2~Tr}j_ܸbl O.ǏCwe>JV_J!F~h(n$S;oyҿ!y96>)Il/7}pw@/|3]d\x}U ce`-]_pm@׻"jO l~ƞ\6HM}/6v/3K-]< qz6ytK-۩}N6 ZJDz<14fB2,<~Jݧmcs=DW5#pῲp148  c68_Qu,i(Q"<26$UŒv,f0e_m̷a߅t&t PAH|PD(d  l)D݅ HЦ& :01J8N KAVJE C )$ 9#HpS:ws/E]gE4AxYFߓ5S,na`KEfê'WkL58W>3CJ_`D$e4ψ@d ўUEd KP1&A1:.D '5qDI0DgD-',AֲU;wWߒQUF~BUOϪv1tDz8lW-Ϧ5Db ?b[]115i"d9\^NnhOg4uP"Qz9⼉'#TLPuE|y}7?6> ĩA+B)DR)SdD r 8s^ <|!g U{򭧥hFi$ѐ@AJR"hJ˜(WJCSzw/ \Ym$xKoJ֛b)cd-8zVLyخM{$A=Rpqyv/|.L^ƫkbaoJbMP@<;3Ps!>ۊ*'0q!L*IDFủ@@P3cn{A eC" .i!%IE«="9r\ȥn 6J/x[ŶVf¢J+myh*|.V~6\9grz%yx9M~8K 2Ɉq3!w\V|eqP:ӕzb đl i| :(VTO4P&ƈ z-uțG-p^Xh72X-' !fEL8eJ<3E>ON+݄|2g>-b'ϭ+)DxGd1:Fυ\-z|Sbr$(B+-\Q\\C~џr)GτU3$BN%FRfGGBR8З>x#̣gN \tCymx1ߟw>o>tG?U?r|u.GQ a?fst\k_W_~~`=L-gZ|W76E_4؋vcW7n\eOY vz2ن[X9L5xٯ?ag:/AwuMQ-/.~ܥ{7uK6(SrCr3ƸVvG'1<B^ErWOo6: Eg]K.Y U]nl WZS2"/nCg oN*)HEZsRDI<12 _̳|qs񥴶ʉQKhab&x<3Y)eqOkrJ )MBk X QY[ _R"z`1T(/bզ_z'dPήpE롽=MUFq/ِЩ)#ά8y٧_ ӯ( ^ YWկN۪6 ~s rpx݀ӊ FM$D J,t [8|&Ԛٖr+z=*&U srqAqW{A(blcWڨfjP3 [F!cir\Evzs13_Ö+]DŽj9"e?׊*W|PFp]PUiIZ8FԮ,"ZJ Vk2"w<.<|.[^f@5d GzPOV{>\sVڪupFYuCbvUQԆL'tE&;s`D/C>1xzFf bTAbc-Vs.Ih5#A20N+ROOwQsaG Ԑ)/&)q`T%X><rN1 ׬Qq7F48BBp[`uM'wl<nK&ЋHʝk1^9c'6|&%ȇvBD6YDm1zTvM~j}>k&( TW Vb3sPr1KpoTQӈ0Az4! #2Z钪BC R,w7wOO d0}rYѰ>yJ֫[N{ Y1҅263 , =/L#q] ucFQsA n`~g'Ddxe[C<.="/=@Bn᫱|ޘs`s)1mT8#2 C>r]q"#x)"tEw[ȸ]C CiHSBLHiҮW[μl758_74g;la"匸AodDF6J͢5W[eߧZݮM&$eRhP[ NRrnO^q.a!i">ރybڽgyf b .ijB<>8enqr/7gJ*6NÚz^'6!Pkn2%Np1(w䶈>`1m. W@82v;%4[NJ! ങ/ޓ_ :B҆"I Hr5{{ \JhHLFJ]kDД1Q$#2Ncx Ǒυܲ1wT7X,%mk-˗=>|YOCԡh,UʦAl6I5$M,*p-YX5%jmCʊ[!z=¸UעPjo~7zrSx_IYIuPPD&IdAΨ/ Qs]Dˡ"1#2dw i+<o_6 P3#2Z.Qζڅ%tf˝컢7am>Sʼn/E8p=(9r\+݇oSF?lxH4у8C"g*8 0\0e:U@voF'*)SS45 ^dy%x᪇;LmYwo6NbVʡgkY̐ˍ0qUXwt6xbDxUumHe?aE>򚵷18'7 e.(ʺ )9D Vkg%򹐫z^I⃗^6]FMc7SZLc#\]d_ H |ά۲ZnrT}><}]+Y0Ƅ4rӔ FUV +bFDhDFñ#*9n.Rd>y:`/ jgU W^4ﮣw$7|.ծƝfaoCv7lˑ'X7 }fȘ~]p]|&rqxsR LjE{ae[QFps&6_'3)w)}QQ"% 0¸?D<1!}I4ɰ$E>tJӢfܸބՂ8k8pd[mvZ\jCd6>EO8U.*Vx@ ߳15q6N(FowDŽ%i#2jn~G"U};b5YN?MiEXZzrD.H ~9xؿq1s!_p˂1MFEFd4=˛{ROggl7Ms:{}5{ U?r|GF'qog>#X?_W_~s~EM+11'^ Y;.]ķ/&UY_ ^B{ `EwGl\=?G}tcAυ(чq"Sb8$i4K+'Wi;ߟl:2iu5\Lmc֜k!Eo`] `.=iN^;))88]p@jbqʍ"G}ciFυbT:!r:cpl~XBRQop4](@7RȈcX " wW]&oݴ + Khhan!HA_cEL-UT%8G O(0k[ܞݷMM6s{mZ-8##&;+F[̴ u=#(?" yE'~\?c&PsahI*zҦs iUx8H8:X_|k9 OԪⰌy"ۭۧү 'Nlr#×#jKz q|I*//'FEZr$*= C3d\b6>VR t7'\ QXO"XzB:o]جj۫D~_k/f\qA>yŬƉ^骊k~M/w}[J@F%WLQP<.YJiH/%h;) l3XJ'#x픍ߕQ0hm !Pɗ|h;_ D~2uu꜓<_F]1 f@[37UZS2AxrJ&|-y*v֬n6D,%i''LzKכ֗9kNFZ8֬0?}5r'81)|0-+ֲs c򳱚>WZxlXJX^cu2;n =>- l"ֆY]0r c[g^ ׿='TA0xFP qHBQ 9Fm<1U_?Ф4a^Muy08ʇ0 c%gr[ cg#ӾJxr.Q".-)T'ă̺߅Qe g B<)1b,z3֗bcl ʹBܥ4lcE 3!7וJ,ڤq)P)[R&q2z!VRue㪓V6n5.esmzHa[bͯ V;ft~5~8ACpqͬ|Vj ksmD Ta9Q(A "Bc}7hQvu۔o>0T*ȸ@CPuA|oW?@68 2$HkctdT| {틟Vy1V:͂9-@Ӿ̗`KX"mY"b%҈1%t^>.x@ăe6yF U }_<k[Comh`?Jה \Y-8q Aƪ4 23-Z(Jv _"G!ˠ4Sh޽`k[:!PkEɸkV69םb( EƐhӊc=w!xǀ-?=+qi3jSzW7X1T1A| 3Hw xc`;*h&RCbU4 G KV퓞%R{'`n6H[ E>@l;"[k \6ǥ$n׍`6Lh!]`BJJ$6l/I [^1Gg}yPwz 'Nb" I;/}Q?StH(4Ha˰tU\l^?n@8p^Dф$Ek-RrCݕuq1]i]dUʢ)ʀ-#N0,.6k8-k @ EcD)ϵi3O51oRܮǒ8-Sgʔ5w';?-a9oϮȲDlndO3f{;\lAkLgʥU8K!#GKޛFM ănG9zOmw Ro#urFNkJcmʏ4YuQ~c]󩛉5⌑dedΦG L|fP Hw3hj_#Mƽ\& ut~1A4gy%I=7K6Gt̍4Cڂu;~8_#HZh)g`0]?z~mkͷKDM+%g @:>zw"J9]PYxfӣmeܿLrhp O=pwtzVVc0SIr9xSѧ]^;Pg׶~1ư)59 I{;}fC8_QX>u{&#C@ +ojxې_2[bBJ+tgܻk\7w)ǢAQRMy-B;Yst2d%Hԓ->=neD*u>_ENcL.z[U٫d2g k|f#^#03ށ^]`k./(j&5Gn#%fRZc.he o9*uݺVN>;w7{[1~ 2\!n.hjJ>& krG+cfwB8l0ahACG{IK@&Œ1S'Kzgui6*qcڶo jή{b s|aHeӧauD1RDc>= qr, ~_XSF*_gB襤)^akAR{ѲDDDs_z⋞/I87m _i!?Y19[+VLB(bV¦{Pef𼛆}{B ҨQo)͇Ҿ̪6VGgm4>!D0pDEB 6L'IƊm^چ:uscl˂ˌ ,t/0͗ٺK2eJ)0zzY) 3އXXT*J~$tK>=!BW!!!̭O+M2G Zxr<9t~Ard\uj>OU=;Үw7eamS:Mwa_ϯ"$ql,(⠱SV'TjR;$Eb%#CLPL27I:QSIibL5a5iZo9+q:P:L8+B &_&O?K3\V_ǡHaC"-K:4'wO8|3[av>T_ϓdtؗ"n;&ztUMe\_ayAWlI7ӒnehzHy%' <׭Ξ/vrUb߇Wlls_07XHޥقU2Ґ|nZd"~ ]]Osojms.&ψvJs=fE D"n~y;;;2֫̎Qex[{cv.l  F1 KEs:O@qSē(QP$ †Emۦ~]I1F1ݵ:DT=-ĽY\=ߩh3_|ok1Ae`3 A 4NGZk)Upc& um26`elQCuP_|=ǵk 3OAW=1ƞixׯӺT/zaGLF{^Q7ه_Hy#B|@{Wz`xz"?]1]"S2 A:ψŎ(rk m`n5ClFHK!A8ij;- rtSrK5MCEPq;!ձ-CHB_'(֦;[2wX!oqX,e"B8tN;v Y9W{ ax<*| ~}~m#=ۨt/]Us4>'d竇iY(bܭ}uDF?N.-ɧ4OuRHqe9c鲾,Mr#uܜEv^g*[:ķٙa3 >P|)r u0RdTAP[ҮrcX2:ʴG 5,nMqR:[י _#9xGqF@xp/*-e+aq^)FQPu& X[':GeDhf;ELu)$s$".SeMLI}mt5 sj)} ϫ2mI*ӧI=O6WM]]46n c:II22nP˦C]83ZeXjnU6ڸ0U_vnﭽ?Es5]Q044O=ڇ0 1A==eQxxۆ:{_1oc*-9rt-w.1؋l?^}[xQse[,~],(. Lݷy-I1V9-v|Km5Z'EvÌlMA^<:6+1K`䔼v(yvx[,"޵6r#0-0a`pdA22A@6I[Yr$ٞE+vzӖxz.ŪbկH* rʎqr(zg4Lk@>^竭 l7g90:Q[4纭AuO9جn#*},u񘕓'߯G}zGA魿(}(Uтi Y)khM?".4y3*ڬ:tLбsj=ʩ|Jj0-k?eǽek:~0R rjM(JGسMwESg^ql/U5MHG.V $Us崾`!>A8] ž1 yH}Mc%7qo=ɯ)PϬ،Vlq +# $$4]4ksLX3NP3[2gW3ɌjSp/ 4w!Hs# `+ @].?Ӻaoe!Νf](Ǵh|nʜ/g.oF'3J/󬰨 ?)>R^μ΂i`w > B{^OnrQ ;.\EmEh,t0ϑ=BV\]鿓]u8!zшxO,3j I~p~M=noܔC_ EP08vqmNBbG F Ӭgܕ&D0I2'(9삖[#*T d>}w nk[a6%)KFEV8BYB5np뙏4Ņ4xk@o'dXA%B_QruvS$h*besvX#uvE:/(E;?3am=MyiD#})n{Uq}nf;qw'UGqKt3C\uEw7' uq D/Q$sSwFDtC'9},9?QLuC@,V.T.-&zW)pޜv;/,yA, 9o3p!T@@M^֯ ^.UAXVr`?zV\89)$,P<R #KqDxasufs5kIJq¸!qoiUS12+IJ H/*;ao.y]lPӾ~ RzfΎ N^xS 5pxx7~8fwar<|4! xRP4R!H$s)=eYjt|u2ҥ-ϖ5l]e|^ 6͡0Sj7(Dwo.Wmg?CoOm-ɖ&C ^l'޵s]Kߐ&ͩy:} #T!$*uǫ_P^}~=@.MO#bRPpa&NFšs9},Ns(rn} }4v%+%H4(Q5cl9+7 i($DHGq@w+#VBj5 rLK}b_0 ;ŞI%_ $qg2u Zaߛ?_:=xvΠz?{{υsq"Q4>R/ 6JHq!=?[B_^OwpмBĀVEiZ[dϟ?û9 CLf-ެ.wW͸w p>ya},9%. lrٿEM 6"ГXYĎ7z)f)8:2#ƘCܚ#As9q9Mnj858atf\ɹfٌkŽ}F'/vT# {虬Dv_`ЄSh՗mO[x U1JxJ?%Y{eK/p$m6)i}qMEpQ$[Q+פ\PV1Rq"JE䰶*jL'ְ,,xleHYE[5?F>AǗ.vt\*P*2˕rnNs4$R֊P!eѠD"hL4cia'ɹ8G~lvW?!Lտ]DuR?gMh]_V_;\Sx Хю/|bZo @;&KcpZ+WbZ7gfQbV<coilŌr:FW@?ͼ:I`t?͝Q}f_Ώ4K\0f2{]o8Si;SRU?& cC^4,_ԡ F> jlkʋzh7;^χMm/ʬadr7zP?(R>M@'o bB7t1 1mC4=]n:uO+N>vJ,Hi3Xqaе()*bZXD#6 gPü-ReDn]X_oQ?kAr%BIsn }=#<2`1cb=0zt{qurgo }mh`>@_e=Zz.ScW3No51)-+MP y$T8{naUlVev?tكy:ջh/$vatp!$3+H>|2G>ĕý Y0߼ /«^g1k!0!eL؊Ɨ| 5Q<2YyUBqڈQ|hF}F=[`OӤiaY(}q:0?_׉oKk 푴W߶K3+vMR(e@/iKS!қL10)x%Ul:OJ[kŋoC65Ÿl8`I~ANuf{1p eb*k~~M M^1_l: ")'fNVv/Lr$Tom&65ٝ㠍Y%ۦr2?=ᓡw!`G98#f!HIQt4f,R>OG+ {%i'[#o;kC59)84 5Xxss'f fށ.V(4խGifJA"xIh|Lca4'gHs2GI3؋ ,e{K'f6yih>c&ΐ•D VpΣ ퟘ)t+xj0rB7{Qˎhr_-K ǩcJN@2U# +b-cь8qPVC EO4ƕrZJ3G/ӌ3iX 1`eHkb:(k힘pt g!_f!ohLQ`,;3CΌ',Qdgwƴe#泀aw A:ψ*ZӧǴWFwC^h l1 ,0OJxj䯈,V>1BBj*DpΊ ,7_!p^ +^{M8p8`mନ(2rQŪJɪ+GP)Wm)E4p`- ra:1C[81s|d;lj+YBr~4ے4lh +1[GMZ*]61|59@V*ā0 P P"Zj\8qB^av'=k]|z7%kyH`DB01"璁}(/y_Φlj;q*XMRڽ&"NE%\Q]vE2~x{A3GccoeQ٢f&L }K'f: sM C(56Ӻ*]o8WbTH 0k;vnH9(+I(DIyvfvL|2q$ TF߭(QK`& #&&Hn}L8BzgpO8 kdS#h.]VSvGA!%rw5e҈GH(IP*2r NImqPLmr@qla'YWCDq>H1JѰQpc(\9T}ɷrHz.[ "xY`0 GIX;1>8?tiyKV{zxHٳP"coqG)7^AL*(ZwI辂,eJ v?;{pR 8#YY =LtHy+klG,f0CfrcRъ7Hs]^5B#b!X"M)ac."M)ȊgԒصWVx3-dMdYnˢn`4I^o"9Ƒ AAD$B$)m:5^iwZmgrh09hGaxϐ88s&e:`Fjzhxݳ 4PN 85"(8yiL$W4<\.*^*ix'd'e'\JTp9g6@EoyFM t[&BiI q @ʴ)ݸC#xH@T&HkEAd*A.NQ7Bk(G4M#!#jw<( (c$KPD(T*JVE=4,ϪjLϓwerhr܎r crk$8Dr~_VIԱK1S*AJV`:tD&6؁_zevʴ%41#Ħh(ύ Sc sC38ŋ(Ƣŝq=eE؁#ڭ)\B/9{%=//~jMlaHAXMZT E2JYbSzhxA!t3/eA-l 's嬘f\WM1jL'IxX1]6xX=4<)vJ6NoD Pa{ۓ!,5ҰTe+)ԈUȨjC@ys>"gE^65GP!1sf %>$Hh-#P45 ƇPc ę^3$dQ;ZQ&m}V9_lpm pf`.rz=$ܷD #2(z;U DN{6$%Xa1C7 PVl 8kΖ!4F)89<$R^g#]* HX;uqx-@+d ǧl7lbwxϨD)* LIvͭ.mלz0yv@ު9p 8VٕK&XĞI=4<#Cjh/r(e(P +ąߧ*fqcUv[ \Nn3'uN,JZ6JOlQTie,D"-#0:I4IJ#$jui+<j5%,"*gbNkg\g/fQZɩ2<􌟇;4O[%{hxǁWKXH] = Ƭ" ݢḡDfH|3HL0UD"^ v/Aq!WIĕ=4gcޯLn7CNX UTHyF%7ìz3(ZӺ%,x狲DKN=A3BT`Ī̈; .jZ2dxِ<@weϰxhrKpҦgp"ca$qΪAYYdŐ e+$t#V=c!$⽕}Q$ø6;X&P,b<<[ٌ(r%i5 h^P噰4ۋBQT#E)M=ABw o^ EBsWǴs*7 FתX;ʜkQnGH 3D!gp}jF'V4qtʔ0 &.4MP7bvZ:ؔtqa)pĪR iQX41a*s+0I3 ot2%(g[E0)6oLÕIXt6-lICfd3ܙ3`Z_y-ogզ0~RGO({ ɹqL>!LcN% 1,-CM6tVYG-LTS|вz3_zm> `%s-x;Zҙ 2=5owCmf{1+v~Ho}ܼ |xdof×Ϯ¡( ePk9O޷.1>/@A5Ri=uCݭ/%aƩg!MXn^L/M,.Kl-J~h:(gծ n/4h9cdne s ]oȭ~Nvkj &|u?v]K V/zQԥpx n%rj?5|~rK/u FVtկNb v9 ,\T5.٢}cMQxW" 8+ckj20Em-TA[12xJ`e3YxjsWP`,]P373NwW0IC? ^Հj-pX`JedFR*nZb܈P$D4 B~tvY Ȁ|y^θe@S} >gNEܹεg7` ;P7ݭߟ;f&xpxpܚцl&i5'K[Q>~u+ }{Zܙ{߹aDvsBF16 7L>Kͯ(al}Od|OWdU1C3rNteX p&(NxSOEF()9af~2/V^cw,`iRf P06->m=BM`ngpNXvFQNxf 7gMw`ѣ`lU%d(fI(JQTr EZe2n9-q7UZ)(vdEm'j!V q;i!.<\G|cju$ϬS"_;MdH:%0 ȰM\vs]/vlX!Y`}Vsb)6jk!ŖiyV>(S2IpfD++ͨJ< tA{ƍlʔd}7#P?xe&U[ׅf‡+ocG!ep*ht>hTpj?ȔJ<vM "SupZ"=Ez0jx3d!'>̇U.0QqDw7CV)ϨC| @[wd0)4{aмNYn7f AxZƊ[%~1AVzTDpsgY'Png`[+Lf';'O>*V9عFD<(E]0ӄ+ljg uRyi8넆BSCiAU)E`)6s [! L`fʃ \J뢣H){z"-E %cxEpA< PXҸ)DD94L)6Ŋ `ݰmhFT3<rZ"Fx`ͷx N(x(ntSInIt=ݝt7,VٚrJ4f>4 Y<>qXco8?MQCl0+-ϳˑ䋍eks .ϑ {9 )Rc|0HGFuwL*͜go{}0cq:(%cH)(liYg`E-{m(DE&bI !^8=5tgt:_v l,B0"J)_u]ݢ2JMnr\毳KS.f'i{Wpl>ASI pV֗!hp􍁕Չ✢L?ݕ^K0W|o#ZV>>2`C7//\DεŖ8 ?Tz-Z*O_P(JC?~jhIP[55CѬu~<)קIUM;.z)ٺh0cpV Zm+FZ&I_ZHl8dj㣯f>wE lRU2C%㡑aqvf@ݫ߼?_xӷoywwx}R]PC(ڡhZ4l*֢iMBmڕ}MhA\[pKmէcr01ib2Szo:qGG-='[kc3|lf v콑cIe_J:nFtLDIR$ ItNn]ҽ{%OP,N)M#/ΣĎl3⤃Y?pwv[%'>Ŀ:W>Q>XG.GRo(Y+GᗇtssLБ %tQBG %tQBGqq61g%sV sVbJjH21g%LsVbJY9+1g%欧Ƅ'欄sVbJY9+1g%ĜTĜ _!;J X9NN RUA0R*ˈK+rARkCX5+TH@ux?γJE)}v6(`g4no`by614?=[( B\`R%5(X{I0ʮ}ǟeWǰV{@W0+L{R)^+h#,e?Ovss6,[<*`XSTT Ղd?FJ`bPA$7?ӯ32{v.yvCYOx3}9}Ҷ =..յ݇e٭RR/U=\pyXˠokk<2sz|;<ꡳ6M91l]V[w-lwb _PY-yCdՒM\:4KPݎ骲Kw1i&Uֻ&q"]&έͮ @vk,ּVY`BPƸ.Ӹ 7nG沥խy_Jf%bmv+u2ۣ٤5YU|wrulVyeץZ*u!md2ϭ)m{0o9Pe]!u7=ׁӆƚ\|dCŸh\ȱ )gW t9zd :2 5>X#s cm(Y_f> DssLg,yϸ=:˜+3Hju''ANFd~ꋠT .>.ƻ R=qg8|E#x8CsDۄyqCAsOU >EƔWxCV|J<S3/LrnM-/U|YoV1^@ˤy~X϶Lo#f"C6@[d'QE#%3FmETTUEeʅ6޹JzTt[~/{e{v߼P|cIhYM:cb'rZ\4pI4sdeXUXon7mFZwlkz=7ENO/w{Z/i~0N(y۝FZc&dk/ǹ42RSR)E` F^Aʕk~3sOaq{Fu˦C(~x5x3,lTBz4dtq?B+|].Qd)Z7!Y mJ<)l >(CV0!}=EfyVc7V7O46i)nTN.(b^Ox!E땯]8Bk:rI?B E땇PLhSHJ+;Y(4V:64@Ȅ Ub0+];6<Ә=2?B R`kS'xITI"J*1ZΣ V<ĭOVԹ$|dlL{{R5 A($X)05llBxm ^eL@L:CDqWpozd~e g\QdpL|F0!G۾\j5_oMr2,,KqI-5!yRn>XxgmgHR[JjF wkw ;>6B<ÛylmHLh4:ΕOGGcW&sAD"G=h NKPi2%!"' JC/Nb%DuB9C $$GkLl:ɔɩ\ʩ\U3GFkeVOwUݽb7a-Q?΋9xaYn'_]\2 0"0Id2pۤs.Έ]H=2G. OB]-X) 0FaXadZ^뤊C= B֨Xz%cc(b܏y2Хz!uEX;[h(@Yy4ZhmUqQI@sZ)jҾEҦ+gazm!b9Z3B ]=.J-k漱9R Ur/HmiSCk3R]i]@ WF|BJD˪4sX*ZXlY^sר:1jٟe]A=GZJ<8ߚnTv,8u?~O;ɻ~~W{_wO@ؓ;>y[87p]sq S&` 7}ж4bhrND=C5-!oӻ`v _;I?~=>?%$]{MU$Wh$gDt$YRBPbK=;=.qqC­G}U[*3%`>8U99ڐ[\L^rFmB2rR탤m6,tۏCkp= Fjm9S*/9mܮP'謫VFzRRD"I?q2APxЦUC&:|u}DX؍fkLK ЅcBB>ZWֶGdW,OS/3+SaXJ d~56/TBij$aI Z& Al]VZL3Z@˂N{ M'Q\P1iꬭQ,Zo=Phql۫`m13!JV,2Ԋȃ7B5_29(2 BjDbVk 0ʔz|!UM%/Y*$MUaVgPhl49/3<ҺIISSI^BZʓ"1KZ(&s\P8)qNߧcPh^=<%6_!6Ǩ\-+f/!{"k HjZOK1Bi{^ oluzҲ-Mˡ~Zl-=Ujw6B'&FL^7nrbRM[=C([$K,*4sYNf]9u~ M;107.Y4-s^= -[Tٞ+z[j7*DVYmB x)w 5Tsꓻg|3ztAZID]_wNy_5RLh0R )8x^"a`ב[ wn)U?.h kjre>}PF K[vn;d>͞Ygdׇ|.,lNg3JoS[J–\OPQ'!(,V8/\_Ilg_0hi27mpǻ\} % S),ɟA2m2Rd $-"Di*Pj@$f =ŕFLhBb Ի)0 IyeR M;D)3s`Ar@"P!ZWR>j QqtQtՋ`X !ZK_ -k<6$ޫ2yĴQ²qGrz:BAP?ʯ$,d:Hs6B VywcZ Қ5W$@9 %#{%ɤ% ISXt:A}2?B od FCڹR P{d~)imsIKg)[3)& NHFKZڶβ.rdu.J6,lY-ʼn <ھ$לoi_qcI!X a XNx{$ sXI#&uP(Qi[3g,R}TWUutFL5l:XAٓHOһLJ,OMKP밯!ԚZZDkJlj1DWP+p"< 5؎ߎ%v:-m@I9`}yzD6fpjm J6NP4{6$ۙfG'ԡwEgC|m_k?GZkUkeԔ5+R|'i˻ԙx+2SXv@k;υ)q&3 tv%f熽Z?)PkI Қ G&Gټ(M ](F YZ:XeVUQ;sDj/W`B).jt;:(ۭ*V!cA] 5i;uZ7v1K78͹,T;'**׬f4Am?ujd&,GXtҡ3.qRa0lHXЁ+ 4$S8c۫U(/z@)G+UXFtJMm : c;3Qx. ezWJ5tM.qq*po먄 *R'""ϜTk0GENXu˥&LI WʬQFYJy>e&v%VYSw+q;XoVyxQoWA ׈eF2tlnBU֯U$~AҝͧGίT,F6Ud.JRX뉲8HSLQ/F>Vsuz(T[@@%^27An)eu;"E RrH` 3âҌt[4  Ay >+9C.83NFNX7c"S nHL "Q;!(woBDt$? A|1ya}b.ϊzָH:5.0Z*>;83Jο  >[AS%/3r߳r4H\~ޕ{X[Mo yu_ Ip ?E]o˔LpgX+0RZD0-꿇(ofpcG9~ll(|85 rJ@A_& y^w+sb+>%{sj$#_-`~s;(QI T˽.[ߚXb>R5MwO\ܗ0*31۽Tm~]u?O_->x]f̍4UAYה-9zy^250ߊ0!Tѓ'wtS bg7eVy "x'`b'%g>7պWdUsH+Yh:>L]jnzKJc7)J:~iTOT46{nx{%(7M7o/1Qo^%P`WA<N' ^x]]Sjt]d55+rERkNύ١[mQݧWcp&דQXO{&X_~ b~[ʩ[4ۣ"=wsȻH5:kG>o'f:Lփ.{M,FͼH-I+t5ZY pT`=`ۇ@'\IImX,?aAd|X> j[ey΢mց#:V5s`0IIÄ,jdp*a)-R孕wAF:0=}4GCc[Y6q7ڛ#/d4{ Yjh>[ 2la(6|& 3tgR? 3g:箥}<%I1蓞  _N_TӆʺPBt*2Nʵ|#m;K]w֨)r蒔ni* yVAƕ͍s2cYT6JA3ŃTg%?oFԿ>L;ӔȻ7fg͜y@_/%P?ڙ^O.=T˼3G4X} l׈` :[6YN J>D7CUJ^Bzؿu@Hm0 P9^sXevv46ΎmlvMwNJysWpܘ6ؖCqpfhCU.` Sm=:ӡQMoۆБ /WډܴMl OG%ͱsqv4j鋵i|=|WmF9 Z?V jZEXvcBcSLLϳb؊KAU@Az 07zG8Ov\k⁜5[B}- )kOs;T@/Xg%寧$)h m-|>[^󑷲y=8OGR{{*3[it& x$ )[X1c2b8hnDmT%=qp5A$0ܬC7 J2hCRt ר]QQ7_> KKB9UaaR"m0s_u^+1|ؒف1W3mgc;?NL5&kB}%gPG Vp).}Ԗx93Y{)C']5k54]r vx%wo%j߻%IPE#KiQ 'ZK(Iz yI"R:" evze_E.qVqf@YGa9AQN^d#6Hc{.N:{A2FHH5pFBb%#1 )DXҠ8KFbRFC2V$Nڔ"ڪY{lk"O٧srd8:~ ]QO%Il\;z\b=^CAlC~AKvktjZ0~*ۘp^]E64z@O͝j7Z^|!Fa^қa:iotzuxVmC<hxq8IߥYToŚ/_{yf"<\p6Bgf/n_,'v1xf{&+NÆkD;j:Rj3fȴ1#Hk#8QG'7i??m힋Q(+l G;Xh7vH뉐:p\U\[.b!;%)lj+EIYF,B0#BXYL^jH1h#2&"XdMs]*6?,?mѵ3w5k_z{c]ƙӘGAc,6ʌ ;lQ1,%! j@CO1g)*ެ-ۻxw#tg=Q&qu zHQ!E=#ItJqTQܦ[vudXDcũʧZ )$46T bI՘eE1p68i <|PkPy7Z$ffj5tcCQv>DYcQ|2O[jL9y,EX9ͧ&*h/9vf1|%gS4Ns!D.%`JIU9 3?,b+@D8b-i=jIźf ѻN'IT@%d>$"ߛ׏\^^/E?~,NG0iZ+Ifv%-LG̺L{CK*5Lᅯ9mvl㚓U7x FCh4zUz߁QwMYÏ]B%gĘ߿ $n91~ j ٨4Pz XeS~Pn i*_+_̸',jG!sԕ1 ,4KWrnǤ* gƽzOoaY%+1˷a™+Lp̺4=?kBn]FS[j*EU&>? ) J#_'eb48R9a062NQ?UӼ'oIvN+DU^S WSs [H_D)E㛮tMѥA/fxa+ IM ר]QQ8 ҒPNbXXH&v˵sWT;Z(֢t(brKMץ93]{-C֮k> dToX(mu_=P`0Ogy (0D%u yIEt$LEt!"'UkCׁ˾" \H☁ )f`:$( rZPBxXG, M#x;!cY^i8XG!1z ",iP΁9KFbR.]:V̚5-kZvւR4X"BVo);SN清j O&כ0&*48W LRh&HscVq|vvx #ztu Yӣ>-%HٝKjMxI6.,-˱l&/U)5Nƅ=ztx"ocp_/?5 +}{"+@S7ozwkx)pofҫpQ>\cWX+BS`)"WctuQLݕNÝ|` ][V@P&tH.5;\4@Xg7 lRaIG?[X| _{cS.#%۾_7o\޼ݍ_*JMEםn:\5$\WHqNrb=|qeFUw?ƌOS-Cu|v»ulw0:9ܙ/aon+RZ7 ߁-I;B5= k{b|Owuݐn4f]Tރ0SB|ﴋ BOW}Wd]5sH+i:a Is&.5}? QE_ި:߻J|T෪z[ǿbw?cן>GLwŽ,P;!YP~yDתET*559OW69~#V^:}pKC/oԥd7O:pZG`ؠwN{Qm{Tվ|(\Gꭃq__.\Ԁ/q?zNE-b^&GBYuBcۇ@'\K 5(ϗqX/!":XFv.;u0xE qi *oT<4r3 3ŷ6鱁db#ЀN 3lVsvJT| 팉LDI=vU_:K\xmPpe}]Y4<0 m*CA ƙARL2BIpN;CpN^DZF6 ,FSэI>pI;^Vʷ=P ,w厯6n/7i+p*^ut&(VYGhX%b3?~!It*:z;%h:R{uJ{sNye}-p:m>ŀ{8l |NJq3J x\$0Uqz]Zh9rJ.!Iw5>O`6̴=\4?Npj38%bO`COiXO^4=?_܃VEDxRy0if/ m0Vm_4O'/n|72q4q<` \ۍ_4W+ǘSE]N̾.Aukύ e˗VS92l%EsE3e %l9ewZ+%sCtC;_4_d׻t`mZ/ˠ}*lMKvZht,͢Or|RTv̧yY6l9dAU鵖/\0w4B2myَ\j/ 0 =P َۋ@k:7], p<E.LHY^ ЪW)pөOjzÀoSZg}oqrzpUDpgaj#zȻ/܏-=yXRuμ)2RM 7~ao]Q`M,ߚK\L77H_=ejv9/ er#?%%ft KnӢ Jтd k,)PV8d^s \G;fBיּ.f3?r 8M(rn5+Ew9lqBrX+H!~fhr2Ұ_oCoI)~܆^BCLA0๴=rC vpm~\tmP:Wmi8z֠'0yyڵ98ywpv/4.YwCO`)i`Wϓ? |!rFBB,ZXJL!0" le؉N(A$0Ey)dA E"BA)nvKp3"5SDcK)b΂wDðqa؉Ɔs%U[ysMtlI&8o7OvڐOfklČ!0#Ti\Ќr,H$uNHoLGI<ȥ ,∹YyQ޹Jb)c~#QtlX1{ߜɿ!پon@⾛. ֌0+x02IQA)3Zv"-&zGٹy콂F%ĺ[˱b݇#[Q`V}|T|;blmf-3YZ(ƛO7d{2^=Ͱ02H E-"R([95$դR勉5OF_~{/,3Z #z!Uxd!R&R/5e$UԂFQ4`pHn"ԗ~`RyˉG]^fN󅹟VϬ:Mp盍7!Nc2Z`(3*h2:ZݢrD`p'%&D|s႞c]}{+z~sJܮ7ɳSv.LRq;V{﹗ʞasPhOb -1%=<CKlSm-Wlx<3t[{L.p8nC=i[DυNdgIei}6M5>1c Afd%vq+ @M/Yy>@l0lZR$n{̯%3>{hJ~*D6<#$Ӆ\ | rrCã*A+>%\oY[ŵ^N&s%Кgf!F Y'1A6'_5t戟mD纭l!j+[:>[GHQ+y'٨<@rJ 0V6|0ىh+'5Dx*eÄW'^1󽤓X)- > LP8NMt4_~ Fbv,zS0Qab`a逸44fԠWZTlhѷ1 rM{Rv v`洈u2J.2ZJ#h)~8G572"ʙQi6g(?PEπ>O\(8 6.X1_{ɱFH(򅖜N)bxЄ^R墲70 Þa؏°oY]ߩ^^fL6_@S,l:1?tqV&NWo iIMXҿ R6MZPKy1}h~Kdv 67M?֎a|{WG DͯEa].O5^s\;? H?/7ZsN݈a8X0kkY֛7/RoZI|VmԔE޾pTKomvya-p:ќ ,﬙{43+fV)g5FϬXY):VU`(Ď,onL5"}ĶczԖڴܷʵݶ5.&֮Ur[!,bvIiIck&"ncJ?@xCoFh+Lmn>Lp: KC_K/|r@eښH_!e>T"yXDY2k@OV7 @4 Za U2+/̫}64Vf݊y'-/*2ѣ=ݚPin΃hkrs!7mGCQaې+Gu=LX;iF|ԱWǬW\ILP+} n:u+C|SQ7SEd)&n5 .b.1]8ѠRRqӻBR <S\y"hai)<1 <0b"B k t ).ۘ=>XD; D%4rO81Ⱆ^@ah qol4f_,?ɲ;n÷PUS]3!3FLh)W۟~‘!o~@Ͳ8O:@ϫ† ^yQqo6gQ * %[89fKku-nQ3Di oPޓrv||j72BIhzoCCjBHΑlwZ)Cފ^{-THƱ`c)U1H*TMȅKZzS̝p|5҆6uaovt7ި~ q+ך'I_>wqKpb&hhF&J7!V $iAa*nz2SKZdB`]H"6*k&Z{A1RHp窥ଶq. CB\P`pt}F+ڈg.A?suaEy|}RX\6rQ X )}^/nC &QX(ɤ"/r*-1 ^MR4D w ˰BomI$&zSPψ50%6ff).AfveضuFR$i"#heA"ޝV(K!x.F!IOBs6:&ĪGLVF*gl:>8Yr]/[0/$?pjt^kT1!+,+.:G繛||sL,Z9H]s$ͤS}юBe>vTlށ3QȦ)(rvqݝ(b{O-ItqzB1C *0\̆pݺ4.3 ~bC8[_6WB_K;x׹)(_ۛoJ9tqq>9TpW\ov8m}ibΈĆ['>S=Œ&Owdz-Mu\x{;4n= l0[zݫOrssFNv\FCC-m/!5=ɪʞ@麮VMδyR,bP'\y{wӅٽL7^Ytu̕ធF!N9$w,|:(v9=AuUtMiUc~u7DatO??~woqW`\>yi0]ݯZ757kآkÚdW59~ه)Tf~ ίOݑFwb'8w3:a'8ldRSΧ= rT]#~ס!},#EgrP Ax(ۯ 'q5 3_Տv)L$UqF3'e&@bR (0RJs[2(1YѼQfQ&=*Uk) KiB9:8hu9ԩKNkz1X04- Nk+x;?U:5_Jt@Gىr#ҤC \/I* t rŽ_Fz;*Agʀ|Qn\ YXr3Ϡu<;߈e^B ЮDPB\Sc\YRaN-┵`vɸ(S:hTu3{Zg .-y8!&ԳF";`X@!kܒDdQD)\f.hT"&AX1AxGAk66-. ѭ2T)6Si9dd覕'? _m1io-}&:@,4aLF%j_p>wQi垍zH_ t5T{#[cV9Tky ֹKm$~ U=iwABY%,%6Y%y%U;2Qs 1Z%QQ:|(.:ḃ:nȔv)v׎.:&d[0!t?>ˤViΝ?;k܈^SY6 9O /DR 0-ZmJǣri.߫nئ1|4 -n&zހͬL MOvn.R~kVfΣ+nlk}֣3)+kNKn構 nWv5ܸͮ[-?DisO`Z'toVm='oh8x]ELiܬ8Hs::Y۶l6+72p:cO,QCÜhŊm{;gޔ֢Qa{.qX@/>5%,x# 5/W8XA~w_F9e":*fML.e 빱1$& ON>1gJ* Ѷ2xͲ鬝jf` !}:vC=[w ǩ!fo@oᥜZ|~5b#u&R-/g0?E#B ZSŤu,B2&AI }򱯤r8#&AFĥTL$1&(řP 82Q1xΒj=k؟l>e@8S¦_jc/4ƱL8/O<|@f>ͦ] "*R`\8H@v^`m6$X"$h¸ya z{!>ٔ6Z,l๴߶́ /=DM ]{U [@ل뛓L"`, n}E6Tq #JBܛ)H3'77p_u+f5P pn]7)H?!L{w_1\5-@@e ! 4/j֤[mK4XI%Cg}q3|{}1G,;=0"ALjM<(Á医F2 ,Z#ՁHOƢ#>'=vL8rcVgQهގ?Uwp|"~e&Km`u4[mMr%NR^rL4x^sLd- +SHNsf)+&,\43 =kVH \;j,f?1QPEsQ}F+l~R\&0{`)O (Dq3`*Õq`u.\NG4H6Fjl8 lӪa7^ ض Ժc<F9WS-MڅD1HG j>pFuUAZN9NgI Q΁T P6J2ǎp/㐢,w .W`ԂHEUΦԢcN3cbuK@4*Y &ΈC>(WUM]߷LJNy4.@1"cN+mxp6,CMɮnt޵6r#ҘK|? sNvd,X`G;UKNywZ DG&Z܊tf[f_z6trSg(@cAOϾ]}}V!ڜt8Qj.DNXRM펮xWӺN\ [@/[Is!syeuN ńGCZO˦a9GՎNmv;ʷf!,e9ݶwzmy=z^jRڧwcsއǣ&o!Cwtrn^L2`F*YHLޘȍ_Qduϥ~<ӖaLy1NUhXΠOx=aydfϲHm~ҒNDN2z(ZykBN}f CQU>q8873u F]A'D˽hг!>z?\rk3ߍ$Α9.5ߌmy>GZq}T!,@T%ӄ R*iArU3pªHCGWV_⮧Lv(ݗ2ݬ!n4NP.2cpAEd@DS^:|ׂaz ut0!S&aMYݙ5(K)µ`8z |A7r6]ܓ\%pdJ0u ޔ:~%7vzvZV/.Rr `eqH-,`-^(|(pV-k MҲwD@4:& 3âҌ;$$" (O  X30E% _u'_)7$H&Bgbah)[woBDdnF.sS}(S:8- HcGQF%&$TJ*b˂ԯ u,M2FP: &ZcЫMl\ YR ZV#7%FwIe[+ Ϳ)P&` 8ƧN;* J;Yrt9+~yxC1hV[w`%:AsN|&oaTaC2yi\ky}li20KN1(veDhq ewfl`dؑ2>S1AG&M ЂZ6ʫb0*h㺼E(`2`J!]Znu '1be _INz/{KHl`:2'B!1qV{tHbaJsGVWPRuۤ'Hzw+Ϗ|ٖ>Cgz PocZ̏Hx! x}֑$u3)4ΰ axhXe`S>=LIkpO봦s\3m7׷o넩7?'ߛ,G?QSNcϿ֭f8r }oE-[[EbMf[`6'`D[6JhnuzFV%פ$(|)%єV2[ ykk)=R~&0abIw6 KEݼM4$ UaeZS۪4JϽ&Ϸ!D!e!x0k5рIDk1lq i譴cwNr1&NQ3' qf%*Ű[)r}q{ `vSvD6vy۳›Y(`uOu2edmwFyIh-J%B))& [)Q[M*Ffd j0D^d,]ȫȈ;%vGa^Tn~yǿ4ZPE# |4 j-$* %^F@HHLBDN>tچ0le_E.qVq1 A0 ('ֆt# Ҙ>$쎋%XW$cZadPg($^2 YobA% 9Ij$f.E!cN,iYҲR^@(viʧ*j|}G<֍;7GJ}+UPԔ D[UP VYO}9~4>ρX`)l0g׼AxwlE^,78<3]͹!YsQokS4͟R۠vEO6N#LgQإ38)4kҷ,Bmד,0IX܌gW !yIzˠKg9PrX3sV.7Ί%U[Eڝ=g!Z7~;VLu /189#qI0@wk'vq,X N[x6m)~.7 x`}*Үpee÷ex T.{bC-G>&ݪxjnasv{y0ĸ@?n#윹2Zd%*~tkw-w)Ľ{Tw|ϜkkX- V̦03WBy:THHtAYE-X9'nO|S+1B\[r.TUzG7?fͰ8i7j^{޶xZv{Nkݚ’.tpm9[\NNV qx;L )ߣ˾kG]3reb1ZHf 0&hD305dƖjheUWUOyBBl1g0ᑈf%>-6.MG{@#hye&ʐѥ *B?1!JhdHq<"?0F>8/Ƴlv%U m$kcV+0ʓ]@OO0ES ".BpF=@($=9t^SeѺT];wzX(@FuLψUWݙ:5 ɴZxYA; >oƏfi^%A)VZHm=RG*߃bQnFb \+j?k)fHF/%|9:b+)ر!,xP#PVvu RbFOh#VF/8΃p&P."<)J@+cqXtYH4rZU%L `6BAo 4pތ~X0a)=)ǂp]#ΊSCDͺb?(c5ϐ.Y3!5̬o=< Jk [SudhǬ١-Jw; aS6k؀">MVsp )~0[o[ʼ`Cx F?lV,OǴ~]s3u @Gx፥Q,0:)[{9aM~)87Z<,Fw;JN)JYQ 4Z֠o77~S^<9?>|EaFdrӉpݐ =a}\9P.'Ѝ -<$:E]RHw[ѩzPaK{=zq@#z`}v=&ÞqU|?}ŠZהsځ'w,sM"U 'GPôcdi]΃G1Ս~f?AZ `*6q̪'[19m[~GpE?Xv=ܵjפMg&=jr`2R[ BVR  z޳j~sPiaR6<`*7߼Gp-4.^qk(b^4ZS$aeg5f@ZI LvXP,KFd(4m6FX2:\07 .F,5*բVӥ)H0r;Qf-$_qIA 4_@Mk-0 ]&V&3'npt"B@\K!"RP>`j?&nzB?F1ӤPp&+=e1UpTZgO=pe6 ?OiD?Nn"ĭl6_o0>.CV~P0Q@ܐFk󷯀BVVP/AuW ޼~]@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J ΙR j7J Xk7Z/%KT-J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%U%^OJ ^G x}o@o^ +E*rpN QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%KV GJ txG 暸/J ֫o] V(J2ZH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D $J QH@%(D r@ևǵzS^jw^^o/.Vu3D :&O0br{#\BpK.zti NMbsmڗrֆo^gVF'%drJ@z12*R֐c8YEQ<݈bS ]yY&~^jGmLoqӦ \??y#`ͥ{szǩs-R+L$;Qp&%j[ɗֻ5+bs͞wU[RhF9Gp߃SMjMQc!-ʻ`|~p1ՒN+_je"J{Sٜ: рxV9)\CQl_:VOYNM9/UJW8"Ѵm"+< #HG,/ǯ8[j3B zQ@mݫL#,e>:G2>d$ОuCKMքFYF*ps vZOeOgǼh3|3R͹้FwSi4fm|*hh 1a)@ֲō:Gu1t.39z#SG/MCDP:^mY]H,NْCCVje$m.?c.ǡ4g8_`@}1 ?B6_U^orqVVUu_ɻju]8G~w>[ x%٭eeo4^Fyw붻7yVXJ];W7#fB7O=66?;~w/z{N);QU8::na;NȀף6h'^6{ xy(;O7>xwbu+(p5/W$kNjZl& \QPϗkܪe77d,s$>FWa Ƨ}ujሇO0oxN{5Q+o'/[2sGfJbJb~ļì%Cg+'$8)n&$'Ǎi&ڽh;J3ZҰO ,9LX3C2hJBzreScp|}¨_ QWܵ 2~H,mZI,g36J])s fD/la#=֐LFw~ժG;NBX9ƠG9Sid{swtWZ_G]"|k2f46O;~'JBd32_JifNcU)7ڴ+$vI*(L8h$3`&(HdRn ǮXGcƬ$wF576)MyԂLH>\_PA&ׄ,;ME`>6h-QZGi^4(6Vi%hUQg,T2Savpԅ)gBEUq׌4QDch!,cF{@挅8xc!9] |.`K%F_y4'dC[ 9=O ZZ%D]Pܐ{DШ&,tEfMk$.MJeD.4\dnl|Ԗhe^ W( z3*gI W2$Ph-Kߌ47Kd})o2lϲ4A;rjQ>S'shG{PRZkA&pnq͡J/:>(81iy{b`ZZ6LsD2V`T#ECBe'F < K%B!^oĺ SkR"o hSdS8}bP}M HH2Cn:l.Ѵ !Z`5cXq6Yƣs*d&A߬J9#] at)oN#FPNj"DWd9Ū|ӆT4lUN*:vw~W&JcoiG^5<UkŻɧE!063` ]/>U'dDpͤLov0x+iXIqڸWjZ5/cE2I6`X'<|YwӃ_'k6qUF֯:dӪ {1Ǥq0BoQSߌ7L4N =T5*ĉV^_Bt 2o?9e?;8@@p_wxM[Kzӥ[,]Gjkons 'JcziLO7_ wqFwwr'2-~V2^ d~U5C{kt%T#t.U,Ya{U54ke*|Hܢu}a틵Ѳ.X6EI``[.'4@P}I!:vnLL@{hYIm$?0G~+PdɆ>IcAedԞy鹉k#N+A)AƼ:ZÉ%GͰԃ(D.:USXE|ЉLrc"ӽM焦zUt%+/M?TPHKnC!$ĵ D*3ƒqWB #Asʺ,+R!oܫ,uʪwk#I18Fqj"$DQj):Kd-AاhpDhȽi4hFEL1VzQKqo]u12T͵gڰ_ ݴ\}9GUj-QzvY(;0ڑ$J[7xx S娾 -.guт$\^z[$&55twҒ?:ܞ%wHnuK/%e2D#QD@NJx4$h59z|0U^YJ'N 2\C/gpB 7qYf~AtМz˿\>O_}>_  5Ň ol[dkͪXLzYZ_Or%0b}Wg(?PeWۢTzRBsopT6V[^ %TiCpE( :cbZVz$$5*ʡ)HIB"i$eBk -y~.ev 4rxk\@L4g IBX!\8@͛Vu]|tr/R"$_@9$TA2UnF(OjrAvu^N##JJgM)D̙kIE hN-m,Q[3Gq,ȸmd5aI0 hm WE4((=p,oZgK:;vw 룩D`L,% .HmI5D|oMzhIRLGmkQo|D!@@lA @F+p, sF$E&10՚&ISܥMQC`/@($8T&y4@$B2/('c8ZaGU5x%5oUE x1 ,DiRIi{82&A>qpHԹAr4B|ȱawڼ?8ୖ6w&9Z'JO*ZH&o#c \BaSöL[tތGU殟8(柯\HtiBBw8ii*Jnv:l(Ԗ G5+%p #q"ʝ#fmZƎP7lY5Ok\Y R6R0(!jb %E Hs3˝^E-`۩<}tWol'W&۲p5M߮{燃7O~uU3/®3Ԝn:t[w:1Yh*^i^wM^: @}%](+?-琤@}a2 ϙ+%Lϻ-I#(H$V^ ,wN QYApEq1CVЀRuZBrc.OR@ %' Jkl6>'*nm8BV4ό9S/ [3ʶ4-z3 M VpGH<(K$K\:SZH4Qi}GoF nJS!mqccxaI]꾰B7V?&3j<&׈j]Gs_gUaQɺZmqz-,#^|ÅBm{ku]xTׅ{բk hT¶Ӵi)r =m@=aq1hAJ4gD%B + 3pkOeƐNr^CJJ*Iyݷvr8 yTJ|_ $NrT*g8P 8m-?l~/=GlNQ^\ntsiĪq#˿"CW"L; ,XLèVZDC$^MG7I7;#e6o ar2գ l3d֗.ǁݲ[=ţq]UiAXg N:bˀ*t٦Y@92-jҝ8x[r}taa.E5shY<ti%3G?$"9nRQwe՗6EH).yBHHR gsS )]Mx"i瑎!F JF\`RxFLSB'*7F+BGN띇 24GU6ˤV>N0C)1۶B[HNۏ ooV *JtBL6nACK8jCB!Rd 4{iSv-PŸ̪Fǎs/k$;L9kOZ<2h'R9XkyP1HL451sD DQ69k5Il-wiK}? !q|}_PW%y&A d` PsB5!KHڣJR&V. #F:*(Q$h>CQ 5PDpJel&F \ʜa}{z;gf#p8-I\8rh<bzthia^rM<'j>ĽNLǿFCK"BbH#ଶq. P M t-ZΖqs<_ /U/xP9OJZZ!Ek97kp׷7Kd}!wʜ+^˽ R䌪siGKLqFbuErAF 7.l9TIp\ LyCu-eLȔw1 I4KKcBJK{tbr~kV63 ONOߚ6:U2&+<+gi^}]aw%f=lt]tZXfz9aQ1{+?Le삓7 9[;@HJƕ@^iּeQlQɟp.f)>n덞ܭٿO< UYjM6<+%F&M C)$/,}2<$F)=Qu?NI%T__": C2}>J?LI gǡdzdx~Dą y/C2RPy)-CVuJ=}hh '9ȖnNk+ As\ g FHE$u -1pf@`gIzYZS^jhAZ :#ѸLØb ڎN+?޶km -"= CtMk|YUQvwF3ϣ) wlyR!X"%$!9 {fPgTZFRӕỔ'6^tDGF 1J1Ħ!Q9VAbFeҙ%r8PDd\r",3LG ңEv1*.rb oZgK:;N>DS3:d$I\{=h^?{WF mc:<"/kffy؁w6EIvOd%bL\L"3"+!l4Tʒh,=u9iY2Lz-?3-HB 1H09IdT`6Ddd@&a0'QjT+\ :?.5: څnJEI FC$/y3D GqP5+,36۴v*AgH)iQ H} 2 K:T:J](VDe+x$&N|T{wY[`d֙Q&5Ù̔w[d:iȶ谕aCxUi};^BKNqC$J< Sl\ 3 6)1À+FæWzo./F㣏F`u,eQRS()I/:$h2kgPM~8~қQޤIU\V$ޔ^iv&+eЎŸǫ$,%٤.Uv|#OZ_po9P→"rA[,@&|fs:>kf>3Qo5I{EqJ`Igx3+T _bq "7[&oz^G9Q٧ؙ¶RmwOlbbꝢzϽ0('8!l6 Oz>Q^6Tn>d{V/_Bl  vQK*; -/7 ԃRX j"cgs'gu0S?*ѧz~y<~c賌blWjW,J!Ju),/uYm2d yV:̽GDE^go9pLU."r&;.1pػvlAdcj^qK댬[|~],vWוMwͯ]ۦuy_̯|#v8L?\A쁺WT\BUG{ʼz|)ɈޭE9^)@R Pe$'!\%+cG>SF(yTV^CQY`ń V5CaUN0 Y#idӒ͍}j+3 ۉjl~;6q7E~Ku,m3%BOYE,]Rx˱)H[+nŜ5rͻuoM뿺RF ~oOG2d ҶշEZՈ@+3A q/kHCЖzKx*f˶y,#2(k&`֊ h`jN q=*ShG_N֬'|7C77 Ц{XhK^| -HpGa/}UUͷGxj1z§)Ft弫XLO!һ_o눟QK;~>:d(D$j~ +:aq칠jmnQ1u(VaVcFpUs.! {<%,?\-*[`Ofmaź9B<!ޱ1[=˲.U$Tů^n[q$ļs6QX>i[5Q!Lĝ7K1 a{)q++YYRUm\ho`0B0iլ::s3d|SRѥѐ^ʥeDГ 3'ASf1[OћMeMd7ԛho=RIS\bĬ+E#OLj窦p^Dh4AfHY0r,ʌ[HD5ej+TeڱwRyEX>WteMҶ,gfAJNQ_a=.' >8bQ;PG"3l\9I#sҪ1NYQ* S.|*z{g3D MnYW?5Y_mnBJs󡖪@?Q0b>D1QL~cVGϳtTuvßOwPS$&5ݽcTVqF;[B}[ E}9ؽ[ٖE$N{=$+oK!~A  /ƣ)ThCFH4"/\ȗbPZH3Wߑ*USdY_8e4:_'3ܽW||:iשf-+o޻ x+WFx)1xgdQiBNIg$hR=( 4KDTU4|иYx\>x`p6Y_Ѹ'47g5sQw#"QYe^e!6P%7 # QNJA2+4%L{5/ԹѢVv_UhEmUTH EGL +yA9B0Dfm@!fr*"trDRGf[JldLk*5ZSn>4S R ܤK;O6Chk>'Sk050sm6`KO'`9= ;$}@IRzl}%ӌs傐ј쬱ܢYɸP҆q [Nj0'b\P7[f0jZi+EbBd[d ,tzZ-F9_KSapY*fѠ-؃ɜBu5gN-"I?HcCSץϾ&?nW#J<КۗnU²(x)PN^hA"KGJoVe d魪<"&\Jah\5%r#M*"a\壶,G;P+iY>r8HQZ(ރBGkH2i .mc5 Ç}rg ć/i(~@&tCT5HzrR8.0p]:t ntv.[CiÒrk\5AkcIlI7N:5Z^nBʢeFI8ŝ c-UwoXNڤul$FB̶yo@Y"1ZPp;˝^]amk4<\Pk8?kذ+f=dXdֶ~A4!5nX3q=]ϴpGwӲZL7`);I ps:zc&HRx% l]`sj{dO6;g[ز'=/|СA+-w\ݹݖ򙛬053Y|xVxY>s4?oԸ-)8XUG0 '!s5v3u5BNx>餏ɐTTTxmRpF'Cb{*s`GϧI@ ].-Ll9D1%b[v5FNCWػP܍^иΈxجmֽ!~z]Yq;ym IBbTw6dqWnSTzsf[q8,_H> 6I*Է߽i I(D¢O 1ȹz3ȦȾ@VP!XNZ5F)`J -i{򜿙#к\@J6 HJ Sk q34ት=[#g?D ]Iwb?fc wRdgPK#O\%>:y) +eWۂjڤgBXEbrPs %Fw?,l9|o`s9[KU 7r93Ǎ4ᣱfIÿ6*bX "$ꈀ_ U]Zz[o߮aBA,^Xz0M]ց(#aqK9uz7 GÚXLmֱζe`ɛ=8UdY0Ƽ%$[,VEV* Չ 5ĚAU+_ TkJ q/UW?YQ\U69y^v>x1y dGQcq|ki/Z۵ܮ;s"9`g\nZ|÷ IQk(SpBQ4Vc ul id`%$BD<X!*+((w0.؆|VA"qb42kP1Km955Δt% ΒԅAPkA(ǣ3ۦ/*Z 9ZgmxDluXbp],@fn (hqc KqE dg%ς&n Ij "D¼3xw.gOS@פ]mv' eլ5'|7'o$`eLoa]4}H@{YUX2E.Ųqgܼ- i񭷻csu ˮO@Q`WmNTyC9``͛jq9`2Fw9`~0RA}$[PJ.& Dr tbBìH5n&xϜy97: b0j#FO./ Ol&\:,9|Hnp=/ϢSؤ=B/AΤsw/3b_gkx7c-ܵM^fEu|YN"㕚]M֚ײ&٪pFuv#3OL[)G plZG[n%EHuP$' %oriLkS0-Rkp"+dFS#!#2sFIJςcp^8[v?.6JfM%! -bJlry9bP(}(M 2['YaVTSF! j440lGT$M߬J91e?ޡS$eɈlN7 䝓ZP':Lyk #6L4k0J('\\awp&NjE l17m'p]`=u9PyA6J^՗`8S>/#!C駻ޝ+H٢|mWʡ.w 9r.BF8ˋ)9Y<G=q_0*?sof{Ubď_՝׳ VSbl6xܺoav9+Np4~Zfo1ER3v$!t[7 ƊaVʯȲR> YZ|4<'z۟x8T,2{C֍ZX#+i7`!y`oq]O7J,4.퍪SMwtpM?nBP}_)㷟|//ӗ_O8S9Z- $U  O\1 [2{8OBU4OEw~RgEπ+aWld*Pқg3)&UХ !}1GTB|?坸 j_촏^S mwa$VSdL FX/mN"؎\iI&Ug$0yE>0gpRggA7*KՀb44I|2`|Le,qB34/=7{mdʦN,ڳGu4& a}nm⳴ ڙPB k);ʵ6:_ڂK'9P vj i|m(6"h'ATXgL! _&˕9e]*̮ik$Wʋ85j/½`:6Egܑ$0*N:X= 5f9Q?j=H&- m՝y~6TGnӉU:::bTcAG&&Qեʞ^!mqoU (I&@0h^z[$&y ]Ulu.&[-{*{Mˡ&oLVc7eLL &cEѤ$R)\B mh:zFE}59#}qfG4O~ՠpaПL;PnxHa.2½geFᕈ٨ablb‚*H$hӆ 0y;9iwaI*7 L@8hd0ZAd$q,h'N:{G0 4'؍! 9jw% G:s2S\FT&&k,/F'Cb{*x ^ `G7A0-d85HK,,'f9D1%b[ lcBq7}{e~<Dw?<ĔJ#P!/甉5Z 8C1ڠM4A +NBakͬ'6сAB/Ӡ:zS?cGN]N]0XwGkR R)B?#,pݽ,w΅ywcw3o'yn, ;@7q? c {Ӛ`+u ;{{P44`4ΈȤ;;\j*h{bTIQ$'e'K>4.Jf} Rbxi%jd5%g9f&w$fՓ`'R\fA>˕G@PL8&R-/g|(W4"t t^A1)4` |@< - >H4]Rn9ax qH]VAU!143882!:s,PAON=WxlY:<9?C}*WNYQE ܾ!kH@s^`mt=qOB&*댷㝢Wk4U*K6l+G(њ |7Goh% DJTWsYBiaڰJTRn%\ :B-"~'ŗQh`\ޓH9wX# ȘЦ@ uu$|H:->WWؕ'nۛAjť1ϬZqq#agY3BC9-s6m35е{cs[m1ޥR6|x4 ȿZ?(t*3{r0}1z#B\f0ȷս罱}wΚu>[Psͮ[Ͻu_qyMhnUW^z.%_LC>EYetvsvgM;Y6U|笁>a4'w;vnt{j ɪmBV5 %7sh6: KB.Zzݕ^ۧל|8|kQoݹ4h"EК`+ HVh<~ N c!:JKvW^j}4Ûym5lFy7BrCAYx&}M GW0I/zi$PAZD:.;6F 8!8v ppAAZ୉em\Oqh__[Y/pjQ. MqgͭJͷ%>ow8o WQ2=q`SF:AEJ8AcL97ll;_u}]{N^ި^yOo蚛Jԉ7=߿tF~|O?|Ӈ7?@oo? 870>>5wם| /Ĕs?Ҍ _O)0'^Wd٫ztx~Đą yC2RPax)-EVuN5xQ=^eigZ@ha|KK+pUWt%%u6=$P\AԹ2u=$uJ!0F:an ]_q=ImPVDTu@Dᢶ @8eq==i`K xhr0Nu)D!)XCFDfZ l{nu|ato/<0jfy&ɱ Z (Q|T.{CG8MA6CE! -ᓼn% YZDP <1 <0buuzXghG:.6JP"7"p))5; D锨ăCx ]vzl=ոW^5I`fHvc}_I|E A5PA=e0z o=2GaCvPQ,Dq , J-WKku]$ք[>o?> s7Fg`D=S8<^(1~v!<3\(a60ce™Ib{Mb`"|L+݆X ;?Ͳ’z8R@yThL!"di/q%X#&6,IB^M F$AKِ\p^CQRţ,ŭsM\6֐b,iigma{0 tڊ`'݊^JĶXsjÎSF {[IWLoxp.,pF Ԣ(ӜY Hv.ĻTRxz}ǒ2&?Zgy}0>sbSi$,P9qHi80JRQk"H^cA -/E&k SʬѠ;$zc<F9WS-M &1cy)ڄ$U{LBRNR$!FbȭNBmp @E+X.ؙqB]ԂHDU&3,^:!b iM`TYgD ܈vf8H@`+kb7L3m%AN E r+qy;w[:@[sfɠ.xt .s\6bJs{ճa(q+W_x4/ C "{p+/X SwynX`C>M!y4z{;{ۯRlSJ-wwԝ4*&A,+0їzŚwX+he%a|]+E\;مTq$ D; F;c8ʒ*][Y #"i:tt{p Z`b0ij$Ey91唗Z)}V^tNy4. 0X p`oΉ%[YЇP0& 8Ӵ{ӪrCp6݂*JAH;IHDY4TA3_I_vy/@ `v@`A,ҲHr_2>ţ%[7[G>h%QUg5~ZUv\WAhWYBMed)~_(٥˱h> m_״~m']߼[ɏ7/m!û Rm!cM8XyR++KnK˴eN$ ߞPm( w8Ya mL"k@jƵYc_DO/?XW&YVN&jQA{5etsץ׏vtAd?eWxeu=Һ=Mgya5gZVrEvn7Mo|BP7tØw,-Gn t꜎D;T9C-BK+hv2-z%H`j;fUG󒆕MWL%,TQrE:dA@"a_tu<ȚN'a@i0!&B-<9%1x0dc:srat>Ex&LzY`wZ" 0E}:]Z;Fę6(7lhoή̡1̴ `V1̖XSnWƚ֒4ZMre0@^X1=uOA*&|az/~N_})E[nrSmyx48BBr;juJm7R5lZ;+$r1i_o?T7@_<Ԛ4?ߑZzJ0)yŭ5!T\$Ţ%e"ҋS' 6dRmJȈ@T/,XJigD0#XOQA CNM{Z;2$2Mc0i`BOFlǥ,He,g|P`(8\&% F9 LIIZk0"ds`mLNC29kS([B.iF-' A#% (Zc1CP'һ=PL" 1ifDh-Z2\tOY"cMG0$DP(B5rYR_H#akWT8]g{x좐%A,i*Leese\,Q zzL^gL r%M%N~.ҞS/ 24%t~q?BfI&ŇT Rp&m>Xh)#?$j25Ƃ "b49$\zɄTijZ2MxUYSfGJ-KḋPJiSa`* 0D"{=6/9H'K1S/i]qHvǫ,34WΏn, |g#Oԙ얄yĕل)HE,@>k&.u)T&<@J+4-/oGdV+h .Cm() =rJhӸ 6&k*)eĕ#Ys$gWA뗽p+ikr56]{k?zݚ[4-H|%h,88" w6q\% PkN> P mo" $]97΃GHO- Mq&,˯r y:u[=QLeR~P>x%^CW6ݻ*Xk󘘓)Q!/: JS&Gm+Ώ$8X uSSΠieϬqCk~+DͧxC>\~4YX];蕊T4R! ++$Ɲ{m6 M(˲2:ĔИM4,8tZI&GRs#2oL(d4FR[*eD2Jp-gW"y+/fDse>1kNJ o!./rC/)O]f%\5C11V~|3ipv$?y #+) }6e`%ɡ`ӈeah lyat\`w&v>IU5z]lQ؜oN7v'EJ]Gr_^zUb/b%a Xobbʏ><rI>S^śI͵Kg'g6>vC)g}bW6=٣*0 VnNRnArS |+ s@mYO#axC}jѢ۫D>8y Z<\ f+';u o7FΎ7|~2E $.uRR&`]~dz`>߫/mzvܵfYߓwKޜEqBLܵg޶Qk\9DtmKFAGhf&X Zɗ)r3-p$ɓ $#u IJUOcP~9 Nʹ^Z a١V\reJ]DC7 mj& - c]mHh]BBs.;!C7(:jYAzɡOrecrE4|k%M|!%mXߪ(ՇBݏ7`lR˾FcSwgT*wʆߕQJ^\~o.nV+vVf*.c:΍GZˏ3*ApIƤ3;ۋ} oGde[rqŀ]\SwKQ 1ۋsokX* #c)x@4 #1\I.('wmmY23mCl$cu8)(ET)MQRI$ GbOuשsNFm>Dz|PճSSJF_\#|9=F iP&l3nzìR3~ag$'I|K5L0lxq1ҟ+?8?~U1M xOGdz'+qdWD>6գ=I;ij>F a6[b٘&TO⧓o=]sy|aqf:mbCwX+gha%^Rnh,/itGx,nOjIKf'ˣ0rLt}?X˯?}_>ra?{8'pBi^[?ֶ֭~[K [EU655 H1)Lf$=aqu=k<aϳNь!$1` !#sd^ziƪԩ٫j{A1=qQL]蒺FjK_"1g g Plj5{Bѭ+쳡D[.dy{NYߛ!:sƩ;tvLu>Yy JFrJdn@RErG5l|w^Q(X-GGrNy%+FJu{-i |ȰYdǃG(c<7dT9o}e9[Խ8bo4R\>3jq=u (z ,^~jkm[ۮ/1'SZǝ.\_[c޳,PMEt)ڜ5qe:?B RDy{t%'s~Qس/iGi?ـΝ^+qOie(. &RsAjCkF`4t{f1 sHck7v4)8,+m\Vq )RƒL)&R7E6h3cVɓqr5kb{KU!u@EDb`$ JdV}/,TJۊ$_a04=)y|·WR!2G: pGB.{+ ;R{ޥ[W1f*IWѝ!e B xAeWP`*A)R1fatve8 H#tK JY)Y8=+Q|r) ^9s9[>Y_NO[~K7X[Y1 6D#AdMZtIu^ov4 &)DBI <8m$ 9[HGbU4;/+m,P~)~PoqNxCT؜iT{E38S62бD/m6UOgt?&hRNҹW哐RjH"Yi,$Nk@$)猕6ʔmfы U"|[%e}>KQX;]Bc2g\NO*"h-oQ=ACX;5:A/Ew2\,LC~H~=F{E,ZɎM,T*KghYTTPkž*ѴRX*JmrqEtGU+wLνSս?tNsܩc~zx[K"76= Row_\6O]fdo sitGlqNwLz5ƧOH#;F~MwO>7Vyepvza͎嚷[c_cwLݓn?CwUo]mCŭvz99!pz!`|\q%c`Zu[nk0V ƪ4c8X5c`Uj`kp2 `#`2 JJfaְ},ĖY 3 ђ9P^KDCk{ M5m{^Ck{ u5ԽPP^Ck{ uq[Ck{ u5ԽP^sn2لa'(m0V ƪX5c`Uj0V ƪX5c`^mΟ6WFӋhBx𩣝tD+5*E`0NJ;p;t1w!G9hf}kN Zw}ÍyR4.&0Fv#]}ðx1k{!Oiֵz^DN;'<0"ĸYd@o#Fl1@VGGI8oMN/'!BX꛷}_eQi|Z{bY]Uq]lԩ5׭mӴviY#9Bl#nE-0Q~!{ W'ںpLyԺz{&\0"ل1ցϢΡ4qֲfn iP7,KS>$ YG͍Y' T |shE$;IAUZ4kHpNHoLT_$E1׶o׀ݩ*mqȖ=38U.o{HƊ3DDj}.$J\GHE ŇA8/ J;nX[(-W&ղ``?Ol0CS4OF*ɋ7̞} ]R&k?[oeinLtu8YyʪNg|3͹kf43,،\,*U_{^;\M0 MMg.&  \Jo=]Х*n*hWZikbYm呁Ճxr xF˂w^9N"[m lzL6K=\N\܊ tkʽ:Գ/aGˎtdjҩ{yt?iCt;^Rjpo{գ^~` SF9#UVHqd7LԈz/҂Nȡk,Y&fKcށrfZTGo&I7{W,3GxO$ gu8RvHt>HXe7YN$ t0 N+[W T8+ų䱛6nK)]2M'ϫN=OOdL =վbt8CLx9[䋨,B+,n1/@ uƝ~[QKb!!n4NP.2cpAEd@DS^:ւa:rNc&y$lk4FK)%kX]fpxRmN'Si__D]%I+Dk3uVhi.7STۧKo݄Z^r{[R-J'xH-,`k2`Dc9ZRMۂ9PI`Z 7/Sڔ}+HFǤ"9X`KFJb,*5ąeL}q] 9EAN\OOjTz2գVήT]~z]~Q1{ bn̗8W׳igDidx7Qc$(n3ݴMCڧѬuERa|['IߎGFOso uV6}ÇOOӇ'؁ipwK^ Pt<`jkj*Ԛ|zwW|yWRO^C7(8wct)#ɽ"?)$3g1BA  ,`~3+:,E $܍X"S v @U][}틍Ѳ.8>٪_*GBZϵ󆃲CBZOXꍤC{WT< G8,+H\90${Sac52ǥ%`^ZR>%sRhUˁ_'S''"5#N]G[SV:.ѫ/@UlGgg#-WMwXlURl0,kwB_(}0Ïy;xF]R-7@*`1Jf C;a0`Ay'5&[pDud !#11# =̓xoo0g[r0~ 3gh,x忓x6o~knV|?7ߛ=h JTrlqv9 f>{*0ioyao9PDDArx_a{vK}@͜`򇕓6hc@>ڈ$!&.vO]ƯrA)_McBq cT^*G@2T%.,2mҒxWU^^^sv,)mΡ؁ IUpHB!jg>cĜE NVFbmr]k~pz|` G6 R,L4:Sd |ba+]?ٻqK-5Lzjծˤ;K2y-L׋˂Iz!l9#-*^mWi);cZ0'gL\iRz1R> 9XbpZzKR+ۼpܤ?nh\BSc;]Hp Ms w{nf и촂"ڻFnpzO犐!D!e!x0k5f,`ZFL Fs+%"[W֝6FN*$((3luу{!'XTELJVsX(vGE`.0 rroDZvIQK"w䪙<IiltrcujL-Z|}۸MN+ǭūSՇlH뇉dWޡ!'MaNYDS Tϙ00P:BPiBwB}8PGZ*:K%bdF/cL&3f[vθYK9 /F:3/=/]>Z.Vi?760ወws0ShddOP8+D{CJGt`ZD"rRV-gm|k҉r ,r!dr4H`@HBDŽ{ Ҙ{1c+˜V!!ց3uS/ H[XA`Irq$5hQŹ@i=ҺDY#Pl`4T,jR'N8PSķCVDw=Έj,_|{5CLHŘ(VJX[(-Wcy`gSǹWDSAӸQ1J WMniN˭{Nu݌ (WGtWOh6ߛR+|q;Z ~Ɣ@/! +I\btaظr1=iCwC 2Ի&eQ{Kf0)+%F&82&jq=auN^UiAXENzWs ö#D˃zWE}ġ;frd~<eh(z"i8;#reCaBi)52e=9rvKLpSj6νF^yIܯX )yujT{p=ڻ,nO:S˲ՎuooJhUd[E-C(~T}ݹ,X Ĉ#ZdؒlٱdjR6Yͮ**ut6V  \|:0Lmv#~yز`w<- ˛Җ| M9熎 t8a޲˯zYVÄW_zD~`A4ʏka)7k] 0]Qwtt1d#(. o`)F"Wǎ/Dt<BO025YG%yJIyO7FOʠUZo ;zng.:1kcTdhS(ZǾqd%E+=ߛݑQIrBL23Af{y Ms kIq{jk^ˍ|ܪ8[C >bZ)W-8geK`|sȮZ=6'D:_u/)pˡ.ya2 J =vԐFGA,{A,Z^P{%(xsy5m_c F.>@[eZS!/޷ 4*'!&cF]Ni皚1$G3Ed`LPɨ-)y5֠|09X]]r4] 1PzdFJx,ZS+TWECsw"2mh<}lyωLϿ䍎^X^ofk*b;2+޴|q䗿ҜՖlƟB\#>nqm=z D*a։✸H@JNquWKzׯ\M߿k:Ӌ&׫Y*\&IFkp/'SF;oz?Rkl@[q#&pkusV]=9'>zaT{2ɍ'Z=VD>U ?|*0Vl0'zz|d-^Jh2=EOVjhOyGPjMO=O1AXͪ,&[xz'6m?^.zvɧ Y;{5jsˇ\g #\kҬ-BZNgԚ2Y4ql:?m4 zщ;/OQ?3z?c~?N}O?Ȍ &֊w; |_]ǡ][tfȧ^ 6!yMG+5fOj _'#Bu[3KVeF埈bB'ɒdϷ2J_{.$_BʏGzW^lkWo&x{|If2;"o~;l}vơTcنz>eaϏ-eϾyU7ZH}$(P̤ PQ%\Flc9D7. WǕӉGWPqgvYW+kwZz\5n|.(e߿|) ב:_64ِSr]昺At6㡡;lmMtޫtӁl1gP(]3iٸ R6ј@m- Uv6'ƶ (+t4eU,P:JI9q`lq2Cڣ#2Ägο[es9/_?NެCߨ]qDG/~CdH)/т}}hvG~Օ %ݱ.cWk& vBI_(=x'Zz&astV=ɺ2G$l3(,6]vPmJ{gc<׺Dύ,Ֆ<><$i|I!M~BksW _ Ov"dݫL?NϾNd|)~q[rCzrZI$ԐT}Q;Fvq};W*?=o1lo~z/Ʃ:w,[qC9wܤ}jmq҇+K=]5Q52 B)gMU`֢ ]!???/d8:*eǦh5DN V/R-!tHQ)P#AB@v)T@h/FQhgO!M Y&l}v%" :8(?Wڔ{t~ٵu_ '^5{zq~o_yV_fB$ɋ>'ӵMˇӷW-U{ ?÷|PESUXM m,Z b@ϡZ2FgOɖbcԜU[w1jZ P*fSX+koR'5gL=f"p0x/ā}!ɉ(.5 bqf3Zl{7t}f0XT5&ǚ1qT. +U;Wktd-m T9k۴>8J0XBpOE}%Bc&;MJp7/Ci#j!c^{BC|HD5 nrM$A2[r(xCJC䨾I6tJjnVD ;ږ6C%Su$l2lj<mO6ffFK{JK;h {{F޻*^S-[YVOCp]̭[4xhLE] H'1]њ*o^66!q C(Y^]W6'[vŸZbLzE5%U$ЊTdOҥOm U-1*I3lJtFgŬE g't9#culg誮ڝf̖o뚍*<9@ec5 쏥xKrL )1x  @L>s;zx)O,&ZXwh"X#hb)U4\RuE P9|]d{:9(&RH51J V (KpXJ6y6G6vB#{[{\[Vm];4_&7QkXf9ju(Gn)ǣ֯uR` yҞ .T V .ANäƐH||N#z+W*|fWZ2:rHL"8X *2SЌ֐f*>F}dk3P %v2$Gi\PlqP|N rR4Uwm3"w?$S嚃'mR|P^J+J1Y2I (1S͚9F0Eɍ;vߟvv\ 5e#T N5ULűx]6jgId,J"KЪTe*4C@e=L523[lgӻlPT7|L<+#Us"g-6^rTQ5NUT4߾_8- ?p;hIX jOVێ Bq-f8B]q'=.ŝ()_j82IzkJ&-A(MCMgZu2򶌎cDZW|w'"m<*.V X4 539:+R fQP\Hچˈlur`*oƷyDFlϗha bM!T ZF xUul?ǀqPX_cˬMN[nyf;D.ܐ0{qfzT.VtI*%q/dc=ThSMhSn;XLж"')r`٘4'xS :8Q6%/'6<>M"z:.}b,}rh3=\>67ZgG"0zioww77/9='7fߠɝZc7ixi_ߢ;7نW7GbϼoUy.ٴovʕt^n쒿1>`doT#161dl$Cns- :@%8M9wհDVb4ɱ@1^cR :dcԚQ!+t2b*[F_$ԡI9}G[zн g|=;itz}T:BT,hN)&^ IJ=x5d67κt*n;8{% vTrʦS|V2lyc:i5s'~Y2csB9EtUPvT޷8I*uUGrֶv\+h_ td1Hq$B2RȇMH:5~S¡l)W=CR$!))V;$ kzz:k%OF{/\lK.[CM8Hkad^oOY/n3l.Fm:6Ixj/I(Yx)+V_Fa390kT U`he)1H+h z s l (D <'y,Bd,h pPT`j("ydB^ Kp3"XgD` 1Y,-%0l\nM l(&f}'~WnNpe`q^ʼXzps|?nΘdPUpA3@GoqKXsmfQ ztwN'OduTx_!ϨjPBLj-MNlߐ6/zD=N\‘IKD'%G"D%}\D<[^hB+UR[1EW Z̷r ;u'č]V\rȲ's C.#q%ԟ㓺1;f[1mdPl$"VFSiD)TVal( Α&]íKJ4gI!siLaR/Z2Y+Dꥦ0+CʘtB6qdlY*poֆ;z2c7[e0m+s3h  |1XzrDYmI`Ύ&-+ȱd%i8$%fa #n1coLrYynLs`fF? #׷FWR Q|WTH0># w~n%"tX1I\LC7IJRk4kV" ,U&q<sH)Q1W\11+XAWI\r4*I{]IJztNǗAc{_=+@=4!῿$(|%TV2[7oS[!IƜ:+Ä{Re u$ BacԀL1GOX4ua< wCZn;[0fvloM߀9ۮ{k_ݻ>޼3cܥOKye8ȬYp))VJTX[)-W&e1Պj}b%)11cLP]JYePv}(Y ;@*.\ШUh zQ/4FШGRtQ/ ШB^h zQ/4FШoK\K\xK^K\V~V%.RY0K^K\x /q%.ą^YbB.{,c /q%.ąb8:"J$GCɔGCx b,)IIqdzLRhvx6'WlS|=}c}p|B[Rtj0OT4_,E~5{p0} jw ,7c7n_NT[zqY]gз 'buןMz2hgu!++/=z?Ʒ:o[U9[VA*7i/ D8/50wy$K[}F!vHgISY9E3WhfQ̢/Ku:}a*JJq0TtX^:Wi'MzԔ6f2l%\p0.Vxtv(.t,$%Tp'W4,N&՛ \fO:k Ւ@ ,Rt{jfpNy'wLu;LT"MstoXiV=X SC6ַ=Hn:ͲW!eMohSHASKe U1)ʞjߘs0z}>\TͨWL8KTbޯ?]w9&+Zr+衯:p׻6ih^O5v~m|r*5-ms48҃umyQ >=JOF[xսg~jNp3x; ;q?h{[$ ܌u gu$Z•-1ˮfHw3u63?T>0j`4_ wӎ_k&V Z!ZxVb#Nь0Ր0󷫑KC*8'wՆ;3ÛJot/`8ѧo>çן>'LԧӇ~/Ҏ"'' xoyD*W޴iMk'M2#w>!u?gfn@~rsп.MO 8ϙm ,P=? j~R랛ܢoR=wgv4 7Z~"qg"x\;/Y;NRVK@0c5 i,pT`=`;m'8\$=jUyH+SpuV3spIIÄa52ǥ%^ZR9lcb@w_{JhSfԥNK+pU_z%))/e+Uc}ၱXiT TV0dHښ9\iW4 '?YTs BX!8ZI ?i/"D-#X6zԠx :+N <R18i5 0Y@SJx-RU63N&K:>!Hg&X_x{q+#2fKeC]ɽir,БZ4Q_ 5*-ǸgqV1CWTbrH8+CI({mzљ̝k$|`?!9;#61&&VQ ?3@cĜE NVFh.5LUYJ{"yy+oC/ep@Caui sImwB^}n~~^}v$)>q |Bz 'ƻ % źgؚ#!!u L*.zA81rEDXJF"|u,@)vGEo_> ̂HB9UaaR"rG"rYEY)4ҼjFwU[%#89vxtd[v h'19nmyˍ͝z:;?5Z-I Ũ=FhsR$ Z) C È2RG=KP GkQZJHA .QrI%bdF/cL#2fge\R^fBb l MZnO6kM;2gZ82ЗK/vKrn )3I-{W=×%R-ƀ Uuu|XcA LHM GXi43grird#b iM`TYgDdIrgFH@`+kb7L3m%AN E r+=qldWNmښU[ ݁d$݁xk_[C :oT:W ~NZ5l8jyrGoON?!~D ~3z=1_}8߿ލڥrBnT7墡LT.zT፜5W=/heG'Ա4wI-^t6[%Ӧ_;ַm<;"nÍnjŸN7G/t͏͹}[2p*S|Ayc[J*l{}ÝWȍ܍^GR9џ/dI^pBS(t.A0%A'Q#6 51ùVkW'H.џKlqYTJ`tQOXC+-g&D+L xhE^תR+@Ҋ\aI9ET`vtX8J(=2D)b2HӚ=*zbF} ) DT 9]>q\'uA-yIg*רbrR IJ%&VRy,#C8(*V'H`w٦ t*AD3M:GbҖ(ZSGViT*W K[9Җnzo<c}⡙Q:˩1y|.q*%L6t((BQ`@Du~/c+#d)V:La0 q7Vo_&uӤ~I f }l#i +DBJT-UCDmQ2+$FQP.T[ Sݕ9 \zU41F&Ts)4)YGQX{o;U$ا{6KgN2f7t=݅w].͂[If t{׹zrdO/kl݌?ַw9ltfl:Ė5m{E͝O#;v~F7vKl1.<)W9|Ca݆?@)QbtYs6YӮngŪm2ԙ8r0{koݪzqlsʼn)UPej 5s9(0`+$אQWZ.]]e*%sTW_,}62HW?]f* 2A3/G5#am'_KjsXu ޒSV# 2i? O|3^>FZ\u3*˾~ZR氵ARj*;f]CO;7$~wc9Xa|DGͬ{38ywtFK4(2b3d*P@1Pr;cTb?lb&ԅ-z; $Rsq+FR͈16&Eg35swѿHD0Xcmf=3Ț7J*M Vߜ㴔BsiÆ"䊸ujN8S'&rT7][pɃ0Yx/ϖ .U'߾ۂ@b7a꼻ؙEV'6J@fVDz;V4ShaP_/uy;⾄ss:m6-u컛kek5`ۼ}<lEէe dMeȒ/v\7l,S(c?mKሯ?YtPkވ1mVٙv\ڦ5ꓹgs%<3#4O3\r,\,wЅ3GK75QKRVKo-wVOz(9pޏ!I*F6!R :6.7{/w( jmyNڨYT?K]ݰ{4%<2sӥB:A ]cX!r3zM7͔H4JɊŒ)Dy ! km;=Y!{wK.]{$ȓf#==x=gu uyߡmCۘ>~6p'r #l/bs֍E5y?f1<'*; W~ݜ,]گ+DW.D/,kbEzad fz[4\Ţ꽷hhBXoIXwu^۠ᄊS)޾T|ٱ獪㰑);V\Hޮ3w&kl3K{0 ]u"u"pZ 1AxGU:X]*9QVrW?KfA͖bXӫwZPؔy:In >{=ojnT ngEn8+v]Y ܪb|mrYú]59WlMrקwLG}|(ߡLSmh|riK8[[ Bjp՚i^[Õ#0IRR})=-%=&qA…9> !2y4 +P-%ŀ9=0"ALj&dIfEy:鉶6ST1r:;s_B9K;xiJa7Y*f2>c-󗥲2ZhՇP nqtĆ$bb<$L8.,-EXA=z-t{CEBc Km`u4[mMrh%NR- Utbl8 qw2jlޮ-G,[a89i$>Iߜ.gNO| )rjq".xwco=m@1u ) C0CȠ0bVRQj*(!tYɅsa=76p*)cP IQq?LIExbg1r6cOW؏pRsRc󇆟C%d?~"[g7G1b )#& S>04u'"R` }LH1&A +KeV =zH@.. lڞ xW@!~ּK\c5I80=@͋vp<*gDdRP9G YM RqS*)*@x UC]UE\U׵>yvCu9vUvwe:_LW7<~Ad~w@PL8&R-gת=ɇAA|0][o#+z 6/y3dd_8b@3Ȳc;?Ų-kTWbQYZ* lHJPr0Jj+Im(ya몓SG^'^dP0|S[u*p2Umdkd2'5i2 $ϥϲ?Ne/ũebGJF?#k'`B8 v,avxQD 9sYځ_Q 6Bޣ7'I2ei0lIiP!e`rbd#D.''t|.OF|Ӌ[y۽\. l4ta_к >1=^}jvWVfv-jP l (.t@ҹwp-x)zΙ%ԏ+zl2O nAh; ]D`xFGB-2D,H#:E&s2IDx@.lgث!z edɁmwFΞgٖ|~2{_ dTJذimu//%vw}IX"( զJcIJSOcPӘfP'~ºW _ 9pVYM h@ji&T2Z־:֦^ճv=iG KapJ9$*C J-6Ro!:Oi d-MF[A:d[>vH/ dLV9 bI1gj!7D"_Js2f-=fAskTn\E?~(jTÁ,>/Wy|cP%c&]-%v^y7? >q3 :j|rL8sV1~ Gy~w0>Gfk_̮^fiU`P%1p\FÏf+ʮ7Ӷ{fҎh'bLvVrL`tiyYeXd3/[+XWZvГ9閇còqV%ϺxMnxV);N ZHH0o&4,4\ζ;=:妖y⟳ dz^}bv+߿?}_}O?|S+o# $Yx M?>ajd_J'ߢ/RB=ȭ ꗺ3Y݀* s{M1N_Ubd.$ )8I,dxcXohlg"m] (Q&K;I/mX輦N)'~ YU"'vIRR %(%qI{)9oMN6uN*38(F=gkS;S6ɂoi,YmmQɒ0z> `W іHFGQQ!:IsԾ{ HٳABFYt"DD)l:#`6cO#aG۲%06Ш._Oݢ3o\Wj8Бਕ[ͬ=RMb@`B18XP IJ"4ɤeTBPQY(ғ372!I݁-N-dV2vza I!t)@Cw'ڭ,S~?]_9ˣjD |`6!S9 }هM_xCħc{ږ* 6ۤhQ:}V tm ^o]︂ŒWg3V˶={-'xK-i`-ěiI[ɕԂ:YVߒkIk4s7#|>yrSƘR$U9W.Y#!A`uIhn]7]Qͨo,DYkd.VAU4pG*Bfl-P=uA UHBQ)^󿍬J$H,7Nii9 Ү-cJ׵K@mT:㓷^\vlEsOE#vjDٱFF56wALZ mA0c2(*iȠ@ e._JBX-CN@gRIeJ#'EvH]w芜~aSuv%Eݱ^T^IÕǛ޺@9#RRh"xhD)(>#@_;ଜMdrn#br md*dAxs Q˾Z ^z$aԾozҚB"V<*A`0 $i ?vO9{czBΈNۗawJZo{ʞLgajMS(>91`Bj}UdA Jh_f2%uqdۙ6󲳿? <+am.~RM5MNF~D!]2gf't*،dET΢(Q; B2#PK*8 <9&x696򉴈Aw,g\o#c iKHΒPBGPHh5t" cjYzD,DAgzȱ_2;.QBv0XSؼu|DK9|CބAE qkԜvZs)>\ءfÆZ|t1}J.V KecaLҊֳT&)%,o|w+XI|4 ՈJRvsT-+- d@`Iј$.?s*IQ6W\ @\=K`Fԁ0WϓVØgIq{b\m1H†bp kgyx2rq9_ث+='{q gv_FFg^`xȤ ‰$0{~K~3BNJt,f:IE4H~f(1+bcim7W %YK͘+ RuD* kРq)Rb*Ip6Wo\1&`lt|i?翞u8U^`/`1Jf 1N-}Y~ ]:9eKR(CxHQnԁ:OAK5B$=\ {^2?9LC@OC:a.csVoze_]x?hF~`.I$Q$*͹c3IT&T$*DeM$Q&DeLIeL՚w N@`ќ%q>$Fm? ))$ qՂ:eNi%D˜v!se͜vY2]˜v.seNi992]˜v.k[˜vY=2]˜v.ƭ 6!r1X˜v.seNi92]˜v.seNi}}By.Ol'b. 1]Z!T0I@XijPݳ}!%4Rn}pX*EVj98W y$Kc #z!MፒBX\&R/5eDlFQ4`pHnwoL5xx ; PXOon!vlU# /r\f^}M!QC$4(q&GIЎ qvyp64*kc&PfTetۨQ1,%{#5}ܘ85x%zzXu:h+{<>ؕ+gy RIO1؊ R!~۪&N~IJ\7c x 1==f1{-#@A} k##4I IU͔ؓ"VFudXDcZ)")$4=g3 ],YW?&LdX>7}E{0K|dy &䈒%>{I\y4IZ= -Pz{o'@Zk>]WF箊%.0+F7؀ϬY[DH(򅖜N)bx$GKkVҵ*u|{xIs/jV;KLR2rjδwFCtm C!fH'Gjd^ wE%KU-T5wx4\>z\?lKPewFP.A`}(t:߆A.|QU_q,ȅ 7x~SL_n@cΜӒ[nw׬: gj_? n2:,<((B Rֺ74نz}~dӢ Jтd k,)PJ.gC ^ouzMC/ڐqϚY"$RXkY0^H\ AF9(Z!@y| P,Fv[6moKk;^ afԸ rǷ*LhLRE0CA;>Kt@PQX5 *X0B`D$/h3z Ab H (A$0EyCM!y L E29A6+(’E*L<%<2eF,Yh6iԘ8AAˢ +.C[05#Z+̌k˟7?1 8ȀJfw 3:S'7&IE.e`G̵ۈNkݑ;]vȆk%˫YBK+moΉ:i9،0*~xuMt)Rj f(42E2&=;7|D}xƠW>@O|E?z#;O_\l3:R f>I nbSI).R~xe]sjGwt6S ?%uJnҸ^\\$(#EfOL_pS]vTA9lȦ|^σ9of]yilnerɤU}~,z,1~f^,ز7YoZ7;;M{ޤ>+cVgMO~8]H=HխT-(9vD)e܇^;k ̊pȳZWV{VP։'syyauW&nYKjӰk7 8X/dCm6.j1OrmrێqQ:v1_w,S M# I˥qO`m$ФE0Mi*Y ޼YePSI)*e8&vKp쾡: <fγ31JGİ3a'ɱ0$ievg7Ȱ#14^OqAS_R;t JiB )Frj{w%,K ]ÛSlf`f_:ZI+`i-SNEpit̓M{ Omzz~_XS;N:(2&: OҦDja1k2`Dm9ZRMDU _FzwÔAQ")9IE$s ̰4N+!ʓH@,5=W 6(B?巫1ky1Pɤ B( S, !%NB5Uvʐ<2}g:])H# g ~F`%LDO1eEjV"2BQbxtLpSWm\ Y2[ҾF_:?Q_? QN/c=>}kxZ1\2m,f:_~ܹ4}*t&T$ϸ>b{>Ek+6i`@IG#L9A'Pԝ0\p&$t1l[  y85rJq%Ӕ0n]<_|Ks(GI|:0>^+^΍)HY|l璘&vl~7y*5)͉Scir 9E݆?`#~è3R&;D^[uKE`Tޙ0G{݋R=مMNnT{=7^gU$(l-]W5CѬũTևO0i*>=ٽtV Z괓UV#|& kjHj֥3d4N ƧiT}\Y?70ۧ|J??ϿۇOaϿ|/0RRJԋENjP_ 짽C`V҉T)E&["i`Y/ kf~+_},xN/.uZ>v[4n7_b`qݧu{f:@xŅ!br\6 MuS@1ylJO2Ut$Hm:Lpll3拄gYG~y:ane}7WaonYl8=#E^?:abvqVn_ q@.0zlEP@%J*2Dm̾/uدa\1ZbȲw%inS3"1]+0Uy'M>_{w AI!synV^~(2۶6''wz~<̿ !ϡE8V.0x߯:@;H͜0DKqL& =6;pi}mT<)}ߓad[J(pE\tKɑ@ǜB `U9{[A01.ghlDMr?j3ngƗAK udr5HmlթfI+T2T]jN'e=p0էi}KPY;G`"hr/sj[O]/OqEC\faWIAS>*0iݸ "_?ݮYymXKQ: )]LNg$E m IARgk5F&ȍsj B%5>CɮV dFP%IΌy<3G){ɅKo!]|>;}~gzU 9Сj.P|L9]I8W`TD 5/v܆VJLj[ 8EmrO9_|!oj*U%:gnN~?\]\r7 N͓ Xwu;%קgsOf<Njn'eM,s /.J|ޝ_̮OoNIbf6d(p}uyV!u.1-il#dR~Z쬾q,"0ڙ~gːb*Hur=]U e||暸ܞKq˓vᜫ./>K!?cWWo_5 R7y#eNOsZjW.%\yEWSQ:۪46ESU1j,IJ1ބm&p-t W'FoՂW[Н  wDZwM7OY*41r5hU&kN$/lS2"Bmf;D ߬0v%)  8P0ii @Z5OOG {x԰nJ ӾlYSgy+VR-_ I3IBj<.M[_ZbIc>5QnjQQsU~ƤA:1$oiG򐲀\s!팪1,bvJF輟9p-u//NK Dl WFYcY-Q[q*.4 rfRLr8b(VJyJX%w)QĘE[ /u+#f`4] GTIŒCgkcɡr"iq5:z;ш)K6x4Ae+4dJUr:"OTt08M1v@1vUR=>bu]u+d󁿑a@FtYn;M%@ Ò9w{ jO}'XM}D5ƘJg 6UYhɤgI9/R'Fj;yaMYK%eچ+Z$)-"S`ԛ⶛9H= dǪ 崊+*f\K-vvǡrpA9s)L,ާ TPI kft=[t)rE 4;MFeLS?xSc ?k⽪Cf9;%TLQYP5hW \~*]+m dȮ+|,Z;Y7s6180ے1}tԋwU^>9E.XSvɹDX+XkU*+fyQZM!3>}/H=?&lSI91ZX8x$%RJ|>jW juZl:'wN6d$/Y1(ڨ `R'RTclm*2:OK☕돗B'"?;q8M0 SfWTL31R]Pc$.Ru:S6 MoFDlxU˧A(vsC a Ġv}j`[qj( ؂ܩɨ ]F}/6:|/q )XCFioID?iHMlBHK#Mo}N;hCyC?_A+ _YU^ron ٠KcmĆ} 6Sm79zoGן^n쑓$ڻ9PO~u滃?a3/..֜5oue80yb/zMy7sçc<28WEnFtܦ|O}|+M ~BWz#ubtXۓR9#6zD7tnܖ}\bG=\ Hbv^ @ձ4 }6eLl.+Ue{+Ck _>͟ޗ%n{gyL豜Ytyɵƿϒs+^kcr_W4}?{4[=6כ/5}1Q^9r@pSI:?`atDN jrtDbc>jfcnz8tj`ҫ%RuN'oXTޜ^^q63Ijs~ߡޓr/Pf@C gET$?{{Km8[-nF;f|e%[cTsC6VTKU5$T)kv]ԍSCɪ2Gy{+TFrB wΥ1ZDk;/Kv3e/!Mn8^0ϼ`=e*L'^_]?jV=Z c 8Uj{<%ŹCyoOW#vA2i5=Da0ws@. <2 ̳Rˌ2R)eC0l/~ؤޝw[dYA)5P88tz^ۨsj[in?/MF8R4r&.WSb>y&0Av,?9S:ΒJu6tc޵q$ۿB;RW[?:up7Ag@ЏjkҊTlz#QTS  kfNGXgda&֚{[`ӑY3sY1,؜|RgqVӳs!nv{ ZB]Oz&jL?v\>kr9MM \d5"0%5WcGLX#1 Zu[3:P14:BzNEoˁ l#= oT9uѪɳ=.蹯^ثկ!m.߮,,dP?>,M )!5˿祉67R: YJHI$!g`=Z'Rئ]QYLI'd0^;A1-h#R6 YUk؏pSgڔl禟@?~Ҵ0nbBBˇ[*peЀ>:M?I"Yt[N66~|IHa4-?t~#=[JV 6aMHstv&kڅj8#q4wɸ߸94n7Mqw^[<{%>OEゃeÂLnG*X"sBi 2x{D"mm`ĕ +nc "EMpBhiY8#IZ2c4"!J[a,".)xByrւp~J ϓjGTrp[ )!3LHF˯.y 0 z!A(2x$MJ`p K=dI(j RB Sy _YKFW n*fzQ,hIES"cm}2t!rӠմҳ>(}^=:*A zHGږ8<,BVK5@(ۗfp_9f)L$ `H`Zq22`$*&RKgl#^03T35~*eڂzjt}]-t1M+OܰHƝn=h+]^I)% pMySBŊ-DxܩKБ"D 5uբ|] *ɪݒ Ch4rRzQH-MKQ@᭶:Ki+G<|ޒhwU{q݃و⨏Ov=U$[U8놥TҼng Y^/e `O~0.7I?R/%5R]4䠚kCؔ$xlwMXZh ~nbGHr6 |_Q "a}ž%-싔B -:aW_!\fsmHy5HdJoco5{r٭)-P)E5nήNɘh+,2G[hO1Z91ZXY_ \:"d dD~;-A΁RJE([*JGYЃ4[҅Pr -!99s8+9X,u=+$H2 -mŹ}ům=DK×^mi]^rL S8@$1}lR9T9$B8,-O2&O `3ӗ0S%%rr##IkAz*:k izRB 0XN vlw3rAYW 4s62TFsUvQI@, RԄYBSi L$UthFg87A̻H͆i9o,#9 P)A*:1F,!yD|.FZy,uf~wqV2jS>~*~2[>?҉^i).?`wnS4czge1^aFOO8;YKwSnǔ'Q8$8aJgX2'P)9 qOx.v3i }nv?Է0ҞѾkqhor+oQ#8K I47nLflycgsZ'iKvuuy|v8Ubv[W ޥ0xݓ^q&'gmw4'o%c=#i#IuH}aT@*pr=*~29X<잛wTuq}+0gLW\hHXY,>$ hNNq9bIwxǧ4D`ûP~~n?~>ъ'z3q"$x뻟154ZXЎ׼_e\S{ƽ>1;U !>~:$}?ƒp5n7mcٹy_!I͏}thM8u/B,X6bKg~W&({WK}qŽu.lI3(Nb(d~m+WDKVmB*Tކ5g< D򤦠QTH>IRRD"pILd^aSƪaSʦV%3Uk%3]ߨz2޿llyrG7MH;xn"4V_Z{FЮ.6Jr+27[HT&hUh!i莐[ߟ{@OJyόT Srt6%-{1$%1ڜcA KPֵE EزdHhH184I(c:9K<8+dmqY|Qɞ.'>sۍ[E3o\- ݚvmC^ ;ebzr rFQ7V"4Q*OFjuF"CJ22\*x08|KRSz|ľNs=-mc QK(Tq-bi"$N Jft Y#rh(a,CoqtDp@χ8輲f~5!ar}ݛW߽O>MN>Oz<>Egg\n~Ɍ%24)EvU+"@zcPZ :a6aŃK,JvA|b[%͓^u2;h{6W_KhNڢ a<䃍++iSfټ**/#]k#a0T"ЬWҊ%@3>,"t"%Q&C`AMY¡Xj]c9Ko0i(3Tmd&x?Ky,63BW ݀P-4;/~kpbw?|>?gFlUbU%WẢ|^! YHd*ePI@neWaEg"@`ShC(h wd2uK1hleĮ&ÈN&q5 j͎5Qjqn*Xe9Op2`A%z(l ^LJ&A _G(QP I@C&${#ǒ|HY%ɨF}e<&xE͏u*#" 8 }38(5E8)50RR_DE@A21RFFF(`ٵ.-0$6 < k@#Y\j%vZ<^u \.n,YEe\.xߌgZW\WLٖ  3C! I319Ƽ.n͎mCnh65Poȕn6Vw B$ڨp 4tj!pgj"ڊ&\L ` S5!SFČDNd9V@G gC/_P>7C|87>!qo!qo!qo87C|〈C|87.C|87mm2D?~-яN]RB_~UB Go~xr?m/_x|K8t}3}O?gzc'<>9x/% &xˏ?MZ7+ӤZgd 2  [=k5qo{<>hGÀD:vOxn0yQ]Vņ9]o[3.7?K /jqt0}gv~6vGcyRy|p~5{#:fs岼3]kOcI+~Fޚ镦LhմJOM.hiF_  ;y"djyd|9饗Pf˷S|wvu. I&{/@fp$C;G̛=!yK[۾`x>گ<|nbxglY]HV{e2l{ ZddhAnYoҖw׿$J`CVgcGI(W:CήzeAD2%fKZƘϝ{.5:-ne&㳳^Y7u ?/^ 8`2 A9Y^} t"{aX;XeT//"ӫQ8znJ RNLJ,tX|q `Kۇ;"=VrHR"zTNX-zW1glM#"o e y>ҧר:/4svfg8/SV ktge/:ͼ0,?`逯F 7KY^_L꬟V7IϦ@ %B]NKtJh:%ΰY,W%a! A1Pf\[yw0}\fXeF})8dʃ3e~paKW>b#!mVڡdzQڧ7p oMϼOV"5#S~lhYz(^ODϹow~7`5xם9_r6<9]̬뙽$ƣ?I޳F!ZZZf{KmͨfZ)*Xe3gbǣDOg;KkJnu5ȶV[*zi=.Ԇ ϟǓT< q,`8Ow+urGeʆۃ48eq/F|~??OG㇟~<3J5ckGuGtNy74j[5=4.G}u]#ג>7fiH_?(i8Mh}\ >f5?_ bޫ?zKyxk=בvt0?* 6jj WvGX2\' Dw߯DI`!"Ij!Oqu ֌ 3.f(lRFs{Vk^3WZ<Id몚x-&I$|LR EqQy1:oCNu4|q(dA̓O{΁.'ʭ[-[ݒ՛Ko:ƘYФ]0P 74|hސRA'P,iqܫ4YeVF[!d >Sa!t%1#ڒht%Rm6S@̞2Ȣ#9D)"f+|9A*f4Q:F)!us?N Pn |;{x} ӴYyλ2tvW  {tBHd舷Vh8x$0؜Od A ֐dI`MĦ(So;(1(=>LVz;eAtHeߏh,86e2(IG̐uiF#zKnh(Hya֔S\_},WH_Ni4qg~=!As4l~'ygo;B}i42jԾ)ޕo;rپl4 =* T;kߓ6>4ȱ"u^jf}x<.k?[=o9m7;|^Yx|ׄfYq]S!7)2Ɠ͜ M^ A ^j>whI;$wqN*e4jxR}!U/t2uT2[YR%5tQv'9I8TveRƌ #Og5%lRVc^g#gp1>AcʬU]tsIzSdn,aԔm;1`d`b"ڞ)ϻה 0*h.FNvĖS JBSRQ5ͣyd[mΖ7&buKHU!_ʻBAE9 Mv|r c71&x-řu @:XLutlB 2x^@מYp[֚Ի_!(& a% aw{~aQE􂇓rGҁrF|1 hlxhwA!ψ\?x@zhkm dru#BtlA+Bg"bS`ӌtV]1P ֓քxrv"N];{5 Q#:1Dye(m﯍G؟&t0ηʖĠuS&T)%E+D.ҧ `5A ̨.e:^s;fK{qVJt%DT3CkeQ Iﲕ98%DY%h($ 2 .**Pyrhy.&Ic=;{]7q~w*}01($,,'R$d;S:1ޚ"r$^T˷"˅zu&։KsN|!ÈF(ҢJWF/4F"CΠVvRo'=cb$4.PFܼIZ &uVH%XAj: G' AżL&H̛vp9ؤR6$WL1 *>Y"%CZƚ[$8# zm#}htu&"|{W1AK:X4JƝ/й"4NLe+uztk0R^$ځk~N"}{ c.̋4Lo HT܊s^$MI_R1]3p ҩP&؜U4A wT@Ӓjj|G1XcP'x$MJkE);S5e k+uBihɋetP8'BU>5A"tQ*dHco8_қ>lT˦qK히j|9C| |nw ?7nOpYIr$Ai_#6IDH`uɺB1!STѦFr" ^y;Q%z{#)R-$E1R2l؍ [mo߼}cY-~&Ɉxkfh׌5_3ú.%E4Z@@ 2M]dC(nq0l S>Cz ,eV1H-JM.9vjl{BHi)kNzY NFRvfYA@,E!Qy/zճz؁/ &f"L*6[|.btЀ*6"VzJNGi\zEy-=+jPHVP&鳰I=uugO$\L_ly> /;`JElMBHM<;]{Vħ/|,#%€U&jSTIɰB{$%LY&oޕM(Mac,%"HC&D6O 9Ձ$κsoYo@+]Yj m)v,_Ȳdxf+FڵNڵh=6r[[)y4=R8[T5SH dYXd:kݟ-ήu.2ict3PY6;I58GK5uM\uM(DWbi  `IFTTN5-o* loUMrZ9c{keő][]ZvOiXmp6ClrqV:)Ʀ(R ht Pq7 m!K4h` =Ka%g=g?9{eT }:Hְ|kbVmڄ3p62(OFy"61DՈ$$"`?s_gѐ1FbN1xidNT\qҀ%Z[H"+C]UnBz`!x3eɼ#Ў`8S'=fK|ے$`}^OpGr|9i}x셉"2i-ecCeٻ8rWsڍl#׹iA˒Z%:bKRTfI);eU(d&~zNttG~t{M+io..N1 G_lEțJ+? d hGJ0ESN0_ƄlTs]czWvg(\8g<,!8by8?os 1cbUdbT$Q4ZEj=XHɻ##4][1J!6&g0QCxlTAOƺ$yF3osnL|~@5wmͳ}_S}DdcꢳMq<]AUv琰_ Bo?^X{7v+,][ ,TȲhӳ->^rb&BxэÓھp^Ӥ=u7m&eqwO>zms˥um>:X0G ޞyqf̈=So=?guvsur7e=*ϸAؙVo[:z_w4`&Z䱻i)|-nIkG鍊&7 iG =zQ #]\]]uۃϿͿɝ\ю5.CS㥏ϽpD ly}XXn"ę: e*m Vu6>[u +/Ջ 6>_m?uu:hc|  L .fcbbC28NfU#'*`CE,ҧm$-VKATNg3.6w^ߋfy6*QW<~m]ogRiͷu] +nu7AfkZ@ !ktUWy΋I@)la/u HN*ruYi5ެVRRĄ[.}@ x.E1\OCg[&΁CRc)ZTZJ%*E̩cQlBIT\Rl/_PyaXmuUWxEL4 ה6&2 RpuH^G\1pF9`95\RլBe M$'ݔ܂' ߑЭgV;߱4.*^>&ntD6GzҤtj:{^ j-\kN: NX (: bCR/t.>pj:%W?TT LLX5[EY ȱpTTPb cdX["qJEc); _k!ɝZ{6:""T"*yɷ5sDwoUzy(>i_9sfJy$1Ԙ3(m$5ȋNg7Yn7]_n<6{.B+T3Tgk\QU]kfQjq.],l0(Vd|PKQYev(KvRZ*Vh;Y7q]5&3Ǚ٪Eme*GrJnŌO쬍:bArN,8RjFN; 4?/~ @\h3gXD% d-"`} H)8AnP+^L dG LgUx7J(^v/b"81aj`-9Pl|u|ןʪ18@Iv k2AW s l L־T,x(@&'Ù}]zmUG'c u]APJU٨2D9J֬.T@ gkjroM H|Ȣ ܱY執$C^~5ވ=0+9i @ff>/.]K(ň% 1g}kT- T]4o]voS]'drj{gRo^8z+T`RCUfhN6U_ѳQ>z|_3٭vSye?턴^i vGZ+*:q3׎cXiV "0z1\sP:Q^y^s 'c7W]zy!6W"e M䟀ȅI&vjЬrQ 7p3_ƴ)yd}\EjZ7sMm27ξX7}-!~w2A>&cڡXYUk*yG(`T0 V޵BKpp2RWwM>x7ٗxyaUfLQHyS=$%2E5<,pfC)9 Nrh`AȥInHR: K8! @ f.&LkŔjزe9;#:#O>|ޠq9gkxzDerل[Q9+/).Ҫ93饴gƁ d4*gsI?脞y^, IΘ(E iμfs 482!k`*q$P kpHBm=}(wN(br'z޾fŽ yx:2 (""YFxL*QlxdsЫzEC>nfi~qLEtFJQ;Y'gm)XGVu~aj0nJn}PJ\F۵~gh|KQ5?G#7.aޏmlP}h9 VXLYl A\&`G&桶l7e-[fal;d\B![LVU Q" 2s.̵LZ:vWm61Hg<$dL*<>#:ѓ:e%g)*+$7C=?9[kssUeby|Ck/~AR}g^M]n 8ВqG$($!A)"H)@HVΪ+|,41vl⦉@ƢўA쎱w#5#6U=@U3TVAQ]P0 CFz%fB2YE *!o$&B \J958؀NIĥ|>j 9ZrP\M \iob+\jfm+CCSJ˕]8;V?}S;4\r$8R#/->sm%Kt&m&}TI+b|M&L.fִ"t^*^Rq"Ԇk $tE,"92,i~{!SUiZxy}郱neMư$OJm)61(B߬4aL*?A_jjĐd*cSs,Μ͖ >roi Ud 8R]C&%sòĈ1p$jY\I^RiT,o,yrt:/֒ _J<_uvB_|ſ?zRV8p5;\L ˘ݳਬaF賛Fz_h\^/#i4`/Ɠw`qt<,R P V& ?8ct:",lNxr&Y~mKIr\`,yGԁ+F+䕸8|jwlǟ鼈>^]N?.Ggz<_CGG[X˷ UZyMWguaER ;Oy3MӵU8qEˣGNlw7ޯ;fw) Jq⾍y2>^ٵőtżmNGdc$%s$t5 b0L6iOt],trѳ1'GK稜=<"FV`&ʦcZYHXFt0>Gze = F{:UtzUTTM'**?ë [] q,2 wFpoHy݆FyArLΘK6>xLZjeCrsi6=+WV bi7<8d,/U\)4eƄG!y9:˼/A&^%_Y Bna8Pb.0jҭ9h}邓I~9+[g蕼#a;hN9jmcVHn9_qr$ W 6{ʷE~W%K &!˒V0EOYIx1&AbѠeh@^ƽnW{NzڴNio֩Cysp/oۺ ?zI\LůŅk&mׅ 9 2d%IG'USODA%!+"H D2hTЎdzOk.#ӖLo€̲ If#o 寘z2ͅ (/ךP{8j:>on-M;$/ d߻NmbG݂a=B:9o[Bgj;VѿB^7ˉ\l#NY8I%)lO]r8q>|aĖ%ː@ȅg]&H.c`>ɘAr(ev nRaDF %s3}!f%JsF]BL36s !ey4[zmV i8Ls `AK09TN$'NLF/"P < WT\"/3$ۂD4lL/M wR̜YF(NڹxXM W@c[Dʈ"k G'&$*Bd8H|&ϠS\kcJ#I5\pGhLIs$,K[yQ{يOkմd[\q8Cϔ@,]ASdS,gFH4 &D.HVhs. \ Vӎwax/9rcEEAܐ:ُIސqސi%;HrUo F4MN{WE`ZqUְUW!\Q(#*񮊸V \fspU\~ʂ@z'V \Uz;p7|xWOpp$q2 WOV{%'I9 JpO3!Z p=•f}m;s9ReǣvcxC3"GR11_Vڱ?OC# =wdňȊ7k.ɂūUM?vh  m/IN`l>N z% "lgs {r|'\w.Fͥ$ŧ2@0j)XFH@'^-{f!7 t6j &,6VKl1d9:,5REKhEw{iTVxf!|Ϊƺs1pD'd+x>6&&VjAz5D5Ƃ)[UK-8"*$QfLYf }U!OΈ&2q.3[Ԣ8v;.44l'>1glْee降Z0#}<{ٗqa>\ZtZ.3~;{_g-\W֑;Ҁ0yY$r.(j$k`( 9%D "]^ 00h4>;M "bPJGa 'BüA!JQ3BbJMtA"#YXp 6RFaJt΃ 1 xv.PM |}jzNf<ү3MЊ: vX{Di~feƁGXtVeLi đ$Һ/P/Üwvg43{+Boh_d i܌˒xȈ,l6\BKx l$ :y O Nӕj/+#ǛRK"Yy,`D՗gU~B׵ $G0bf-C%C $RڳR &i7#;kO﹟]^12-{2k˟j6B̜EXNDH07~|l5orwRoKS,O|a>Wjw9.X 5 Z%;mB_#ܕ_.j*-ӡ-O|D.|Z%H%YۿP'?> sJVN .Oٺk6 5XgЄvAb1A2 9.&v $a&W5nNW{Ӛnmyc}QU3~bz>M`Le׈7SJ4>IJ;&+sF4~sSˋ~*ԡsa8Y`W"(il"IA XƝ/ts><;Owzb/t'YĬQ,W`$i,M6"R*(oۄ(Lbk77*IE!@%ɌB("ƌ&}ALr*Ati8d9;aqyz,7vu!H)#(o a?+x)=[%6rV nDc**˄ &Ex DFj& RBmPg4 ²uɨQԴ!Y( D"JP"Ɣ.Qo1@kg\A-- ![HRdEb1)BR RJJYޘV"N 44ϭQn]c .39՟:?O@nsgi99tR`1_op[`[jwu]n~&HN=kvlLEigSFs 5$S+H-8ȗRA4\/O$vc J@2[փ`3W\obMhY.}X?%$\ܮr+]OS#kz={w^~iQZ99\]΍U:Z-7vA{1OtqtC)?LnZ}ő]]/xI̾&Vw-i<œnG&ُWg߳ hi98t:4&a6Ug6YYOSz7cNOܜ_d;ɡQU̍&kVR>>|_irV&4停-ˋzE'0=?fuߟ;_>*G?8'pI",~_ ?>ajho5fjy͸-{|F{gv 7 ?L~ 4CM^+uf\P]y_aO}g`y^?ߕJ y3)ԇ1_0;^6R6ۧ qÃm.$Ili Hs 3! ഐ%`+.v\iW{xN{xnO?yzAy0cM׌{h՘uvBP}ꥬ&û_t6~^Y4>Y2t1#l).;cMTc=J[wHh)h;ka-Vr $%K* O{E1fS$*KVs%uS2ɢ3y"I15FgrȍyR3qȰHTٱ@iʽ&6Kjkgp_,{ }v}cȎ;"]./Eo#KzKCSmғ;*. ۑ-fMo{h6mk;cSKI9Ab.EZS4tV.8HўPr+#/ H^9:s0^KYe3߃7@zuMuDm񳅪\.E:+.T`h@V>3vL)NfRt)~=`nݣzv//$RĤmhʶ8kQtTD2+YtmsKO-*D!1$UBɚ9u6; `3q x0}MmeշQ4T(AHMU5d!h! ڸ(8S|HmFj7c@%4?@F╶:#]Z[Nj>v˴.jtqz_ [/b:26_ 8X'/bly:+V+(ҟ)*oӶSQe d :[4v5>A)$}" Exh682DJ12IJkRajyqfXL3v ㈅/hi"2*؋} ? ]:99t2}bbض&c $u*Gm-IEJc%Tے:b1Tg/:aS;(;};bQ&>4FfxLwWMx*0 "6ӏ]Q6FD9"∈iQlRh+,j C. jYҐ!~m,mQ pްoLhR7Tȃ5'2lxuMpk]8#cWl%njx,;z5t%Նc(gh,Ixv J@QSQ`ŗ}fڱ/xwGzf/sVEW~ [QI[$ُ_+1(Ps*p0o9^7{s*tcso96R2FnCv |ߨ*l'rǞqb&+9h©~$yo9ONN|GXR.&~eՒ!`-S@V#Bd0][D *;!]QM]DA=GB$ l7geQ)\ gOY{2" v$ Q̚tA$+SnI[|IAFKKq&N s6 . (m>"Yh32Sf?= n`b^( j%ͬrP>FBD&^R0ҧm꟯)v56T|!iM=#)x!!JF#E}[ oY]z0h\xՋqMg}z~ <jkOV)j|~UQejDU3`^,?v'4xZ,UAky$CpG&|0$)ay:]brbHo#X.B)\t)&kɘ́O*r\ͪ@UʴTHxOU~.pLwBz<\0 Q Vz3؎iC 1>}Y1b@sѺLkXgɸFB|HؗDmpg.xo$@lZ] )fb`^)Ϩ,gkI3_hjL-3 eZ}Ka o4S:)>GLR ;!j1k> *FP%%tf}v`GmfݲjςG;;D%=}IY6 zVbĄd$"x:6$J^i墐ʇ4;?{WF <"0^n`0<B),l,Qrٰ%RyUDeFF0P|.9I!־&vVdX*`%XPq t^ v^׉KsN2T#h*VC6RFW-@LL ‚q(KeCNG*цL@mpj._}jELDdהo p*21h}_,Z)EoWrJ omB!ʨ K\ Qt[r{{\Z1#Ɖ(EEk[͟;uC:b|>]˫>[):-4A@,q!9s}G;ը&EW_1iluPIf9%H1{,@EK Eg5(S8@Qp$B(w;Ζ,Ԝ?y (gǟ杸9v<f4{ ])7|z4+~8ɏ[E(]B1*(¨ > VLp)L2[%b i d%b.^J>#1.JW/VpHLNEbl"ՀIΩU~x;Z&?Ԝ︄%B|Ԯsqx=,(E}-Rl3-O1_pcd A;'eW)dAw)zpLL 8k!:pƕR.T, .0֠G+F)kL]o$vZͿFGњw[zrv rX\֘Udn0o. E6\+cm\AWΤnvm)}4m>gu]zɩFk4WPrp'7/ãVAZu~\OnpxVi=l` TFqK ;?l_U$t2lMb_Ts:XpS,,75rS,75rgyId.{=n!5~%!:`YᯯWඐwo^m3W]7,|-x#4]?j9}_Ft9 cu޿^w^o7o{4jim;Fg'/ޮ_Ξ= XyǎҀI&3k+x0q6ʗDr}UJ;V՗o" G}MymHF;I'yN"Wt$"C2&+$ns4j@$Y]C .5a޲zY<<7HW2ҚP&yzy-W)yS i%mY()@#J!P/?v{OXb6~c$^Y4z+ j(r-dqI0 }k4Ƭ2BόHٰ0yM=҆ e->h,٣RQY(YBbL˅R>۔ *+HL{(, D91$U9{[$wJ֋^x3;[2iwhp;h@|r>7UY&Y2úޫשdFR%-dfHjŸYM|֫z1m@'4~)$Ad%Jc $M.uUw6+*{ K鬉Xj/ՙ_**yqg_Z5lZLNSïW9J6Ǔh{sIԩU"**Pt0^%E BY HhNJhRЙO8[r.1mEPrɡGeW KQR"FkYw6[Jg3ɸ-B¯j =cWd&ja $oY& d-6V_y&E1`2$œD̎EְtJtOC%{!+cTZUep d[oan'/ <7%Qɺs[v՝;GybL:ijv`H0:tJeaX:JP""L, @ :tj !QDF[&?{WFOw/ŷ;8`ubtq,xYbKeGNN"QT*>EV=2*+O T c*& b:;8ީ_Qq(xE--hwxd4RRh# ȐVI(F HB)7pP ^ ao huR'T;)YtDI+`t5٣̏:]wKYgUXE5.Zvѕ,FCN.aIcPWSR: "vvq(=ԏ0a7鑛_GKVv#%rFA7F?>Scc0Fn!e[^oT~D1xDc V(A Hx7 A$(Y#RIM'cئ@]DnL nQζ$JHRpکhҢ څARFO'C1F%f9)7gK%)trqN_2TA&HQ;Gқ" ?vcY% uN@ ~z GS,ĢJ!\`2i|,9_I6Fsg4>sBO鍰YE]%2iFr!_1'wLcJh t72'0vS1$B<ɱ-3)iB5ZF{D~ϠkOXy锥 wà8y7?d5KUR>@9#b!xD1C1$AxGkmS06fV1chð'V! ;BfT%̺ᅋw@fJMZ3vr*C`0 D@ֱ#m耐M/eYGjӭ-*{{1ucUnx79z)!S E+DeէI!+j %XM-*}5)( Ek̨4[\> 4)amcc?*^M5Q"[+C&wʔx9S&ʶM EJl2+T. 'T-m0)O}19??}+edQ[`RziJ{p(goV #~k!U 6&,^wݎ/Vȝ'!pbB~,Po{Unq~uh{޴y|lrˎ[v~m;(Gk-wdջ=1"%8a8Pmm;ΛnW%o,&.F|s܍oSfR} ,?7,r?j¯>s~٦;Ӈnyo=tWEip3bʪ-RWhrP)׾XK2G [@^Bi,#^J"d!QB'xҥ08fbi{doUbp׶軷|&T+U]l/x^򬃋Kb@v{E5?+Zpu/zmGZ2u`1 Ed&*9'J  )iSҶ[eԺ2BIc"}-P.Rd{OEG9.uV<|-y n[sݜWts/ČXET Zm=ڼ 癠;t y(_g\ h5:/ VY)`@=6GA8V8ʐ(1r(2{mF- H^6AAіRH2)QiU!J`M+kPaQ,'Ҫ H 8 #9$cu%[3`r}c|/nlW?Շ?aIevY>L96]VgM%b0STaq.CEGmKtN飈dx7y h4j;<`ᶜmY>'u-rrֱ tbI}MѹyWͧ^yq軗x~EovَH뙐 Z)HYK]ח"gWpvnZ&1B!D+; E!^pkLSk+%%rPJS9F(BLHDNGibӓ:*d-EKra U*Y$, XDh|RL=Sg&4+:t:;^ vlM3t7R/ڬ5AڕOIԃ\.X"LAR@% aQ$[-Q$i ;clRPbր>C1Ȏ^|(SloFVB57W3q8q͂.8n{5&JS,2rgLƶNT ~G-U-,lK.NZ8zC=&V(Mz\_9@%&1Rhfɹl1>M"Bww,++A-3j;#O^֋;1 b2J *5/_ L6ءsvmn8`瀙\_[Y:8ve'`%LDo| 8 7Cn+ gCǸRY ބHEJ̾DO[T8&d iI>Mr) 8uUVqqwU,rWv)O?w7i#{_lt߮>x7d8^//"j(:+Q fYdibl+yЀhhbZZ0|O#HdK:CXs LقpyMvLmzSb@?s}>9#yzGYnso.Qe_~q^_%hU +!a(L4&XBb(|b4c-󅤎ڰlg|2:S":bwI²O䭧"h%i iby*\G^>h2l9^K΋œgA^>˳w>&gR|$FooQֻb?Z @Qk<>,*B6K-xUe7_%(B'W /K։d E6XJP`4;o#Rl9 K>zdَg~tQce> ]Mf_j5Igf8UOf\);yw5NkH['v-U"zRf'4SdB VNGL M$^Lߣ|W*krqN_䚟.[ENFN?-e^8^48wFr@*"Ө]ef]4]d짳w?wk_=?v6Y3;KZ (x~Χe5Xc-2iu! U~Õ"^T*O?wU3Oݧ{4BKk]/(|uzjd-XeU-W];ۤwcLeF_} gU߼vڞ3yUxtأu6E*gfar*dgB d V4'({5H 8=,dFFF s??2Hά$C1 hm| $q]g*Agt)< qm 3(UjSGGfgb6p̽g~v>}Yhv K؀)ԕ@%M ]RdY‡rYxe ?j(Yn{)IjGfJyQbE9;.ↂ*$=y>,cI-2*+&EWm0Jdui qODzo_A6 !ۻ}T{x_-|"G# W]ndGrlxJ=# ֌ 3`I/6!W5=l#bUQ˭L0^Xˍ̌:u*/`N::jG.徭3y{?(:~wpqvzkCoBgBc: st?joPuf_P鵣4GHP 4QGp!WmPQ&7W$ho 7pኣV-~5Sī >R~zr|r-=f} kVw(S+vAgr׿-^ONΠy=MJm}=wOWgkb){g/ n΁N9Կo%|/W72a["Z}.<6֔g '<:Y\wpUkGHݵ| wTg1E @]-u&yYjny۬ ,]̹2o/,ghPO@e):Ub@Yuxhz_?,ïWެd⢷Wksou9uU*Zr|MD. uV:An>fgu}j%&SrqU-х:WkUnK빔fjMU >G;MSh~9k۝ѺDZrahcwd!Etalis]48MQr9b1a2Zt-ٔ{j"׳x_ri k+Ɔ8ZE!ko %5FJH-tϷzn-xLf 0&hd385$-Ռ Ӯ:QK*!<3OՓ&[DBTR=G+(l2 ١L6(4b *c1m m* Q),^WTr$~Ɋ"M-5NY^djG2!hUUGpCΩRRnm5l"5BȍK6JW3Bx1щÚݜBOZ#w:>`Lseߠg16c@FUy*iML~`9 F5"C buS&AZRMjkTtE: ]ŜG6T1K,1 TX&}g ԅqA 1 ٞuTV@N@Q9^ F+USw!9 XP[=JNwp =yJ6 hۧy㒲q4 0Q]̓!W}(qWbA[,X]ql5M`1uV4`ƅu ; YZz|S l}n&U>(͛հQ"AMET6P[ Ze3pydST^b8gB*ۑ+wr9Ȇ2I4NӜGD ʕF(]鬇-`,3(j g7d2Xm 켌@s| j3tZwf A&'@ 3X "aY(ei89_a tk)&@A'Y' sGQ jb ;r- 9oL6 _@k`Av*d*x(Q(ʨqB g,Eyd(_u: y i4/U((1*.oJfRR}k . f1J_5@!ѿ( NRDāf`3h5!! H"9Q dLj*T+AYI t&iP/2u^,KT;f%' NJ1K=Ik!`BK>} ̛y\`_,nт[g .FoG]. m%f3x:p4|K#ߕd+e @Ue84H (v5!N#ė5XH肼AV24ETLԴPyM,|Q1w$ W4yqt@sDAYIgk Vx ;P c}BY,TG7CW2=!W1DҶCMZtYI; >I vz'P'kWG*Uȕ'X1of2(ccP.u%6hA/f9H@%_1f;E42{Ezgsu ֬sdž@ 蠀5HG7P!VvuL(Y9(#VF/$1p$ߔl% cq:XTԙ$@fZvZU7JK)2@!~Ѓ@!o 4hތ'X(C0 tȳP(!cK9QS|ιJV=ot'XwHgYtggNUjth@YL6Rך󣷦:p$_`A-CUvP¦l҈ ໅? Wd5W Ǜ*~0[oK9`#;ϛNjCǴ|@`n`ej'EtuYpl)"m|c2CwgUGŨվ[s55Sf q5J 2Ș3rtWӣ[J3Gh"Xhgr5wn+*X8H`] HЙ &GPz,=q@>ۯnĊbPDvRA'w äsQgK*>z@-L.~Ba ˢ u;b$U7bU/`-й *6qL-6z\9X[1srL@zhYXo8FMg&= d *_@%'-9 V{F[f~RM#bѫx``ELNT09 XvAZ #C+BjSQ AOj|d$NҳjO^ 0 JĒ # t 2x|@+:T6z"frXU[.!b!XTZIoJxI< ]Xvk5c $Ti`MHE@޹F!J L0P5Kx!קl]lRPȕ^$Bc4QQ￿yʞL-ӂFsRhڎf?l/>o1ok\?l/1'xapZ-:0ZH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 v@?߫3Kf@8o} N D)'.M'P qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8ޮM}r! p)xثw!J @N'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q:i=r!`ȥqqVuL r"q!'Q˙@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N7ti=$KуwߟRSnry}]\^P hw]Գӫ0hO%LzK7Ž1.i }%DiނqWC O<}J||z|<_\rZe؋w!~xx~.wHŶ-?[};yV/B,((k0x|n{kP}˽>1=.<'?7ÿz& D_j  r1F~@_귫a+W8Q5|_NMRxDO ub!"H%1T0DuV[_.Rq}˟{"sgC?.Koy{ܗK@ //?cLVQcܛz[D$-[ &jv񓢿k*SWW[~}61kkpY[wg SO񅐫a/np׷d}~nc!<>9yP2}wq=e|Dd|w\G2/r|tDptR,^vɷX"/eݼs^/ΎoS%`Fa$DhI>FMՔ&fl\P7_r|qYo|(؟YLZpb\1,_Q7{/W^~nAUN??wljqٴVop|O?JHt Y- U5FB):p i5/޺DMsy  / 4㩌 R 2pG%1r)6[ì03l< WZ"g6ʋO?^OUgۖ؁w5<ۧ#«iwm$G OCCqy7zB?%)R!)!AQ)RTK4`4{U]5ot:0JV6y3ic֝Ӛz"D[<}I+Lg>֊t!=3&D4#^C:O2"eB^p 4pgp݅Q8%oS ɡ $`[d]LV]hNU/׆ӯa֛b n?J 6 PTZbü ^*!3wv$|W.v͍%LχcKލ' iӎ%[<rz2R!)Ye0bNjp<?*ɤ;lOL/r:]A_Kգ[ceN4s2;Drv_II&v?ԔSZjiBh ˞3'2X_ U2_Z}ڋO̝N&y}\fu\$umcz4߳03IgOLB-3=uRͽclC7fiS)rv t({ Zl h `z Usǫ7ݪ vz,hMLfh eBhH0BO0 vi!#H ALﶧ֨3;L+ x.(nː&MF2*rTj6Kc2҅11hu+kR]b hSds^NUSB雔FijD@a"v5\G6$CRrf2S'4Q2~NS80RYF)QCd689rb5Hğ" ՗LȦ8]qk0627fc%q9` M~A(s\ՙG|,97MVr8]fu">TC5ta& OĀTI&'\܏9l=E W)XO] \[@ WrAnu:>={M}JՇe" }nSG᪺piZ4u[_#mGus$mi r@r:y0=Qm1u6Ϧ51Q>yx c» }hxz֮,+dƓYO[,Ư!kFRm)# zuðXvY6x'\GWBϮLN+GeQIu\)F\YC!m՟&ӐA5IUGt@IkujzO Άqg߽>ɛߟPfNۓ;\9<kI0&G`#3Mq~hnC[Vr݂o3.95fJ2; a ?T?^~y3 g!{^>k͍Gh>3ۻW x"WJF*t]4T6Us/F!_@rd=:h.^_(( 6(617ătݷ#VSdL FT/m.ˆX;p}d"H+q҃Ɔ^ k @sx@eP&Q X$K KM^y)rWNNN717;{P^skcg:ێK/?nPKs݆Zy*uj-Rm1DR Svp>_ᩴuA%\vQ^sYekj'x>4~>O<ׂ;ۮq`t8ˠ#EzpFQEh>i"TKodNCG$'nJ.vcR ؒkYtJO]޲Ds@2R(?9mSR$*ɭ5=ԡf[WYa/gP]Wsy)8P 33 ͠n~}ۆu}7O>X|2^dۛ>h\7oQj=[&=CAd8֗AN˵LZEmgVe[Fx[BP C+wBObD{ {?CLc~) f7/j} OM^ٳ߹˄K O9GnL=Zkx#r?Zߙ_NdzVۿP!/@*ثmEvԣA\N^L'١߻zL/GQn7j23N h30ZkLqP*$!fϓň,{Y,"Y6DN8%qy2I &UPD٨CPiM*VpJ&fZQF %Չ"Q4:XJ<*9[2+)݅%ٷ# 4i#TWC ( Ј -%$(%Lk eі$@&Ep,#h%7FPeZJ9"g3J)j0}*O~ILUV!/{r(~8^ MOMYפ˾yxL*oSP5Bd#NYswG&N<0F G>E%2:Ř'NG!1YϢhRNCDRBd,FfXvƮ ҢD2#5Yۇͭ戕,t0mPhy4|܇,eG2.1LZ' 2&*BR!je`e3rÓeʞpsUW\iG^P#&&,RD6%v1r6Kl?%so;J$%iWR[AJ \,X*yYesP3ƣNL/BP@ந!2YDhzMl{C QZn RpjҔG-XrGT )9PfY3sؗ"gD|=Y_^tlvɮr rODE/h\@$kYhO hT0\8ŧ"}|7y"lCxsFn#z5Z7g{ ' 9k ]*ǺhʏdF4_#tȚQKks {^~D $-DHjʘrKYs2Z79eHe!g#BAnbZ$$5*i˵*|$EcPEf- ťb2CqLe,<:P#MڥBs* NRAH۲>x))J!VQBBiDžg4BmrkO>&'Yv3}tvM# ,s=\ș4=ik1 ?""1\ K2R>';աȞFRDfvV"vc83N:5J^i"EypBX!;Ŝ #B^׾"[,]KW2Ndys Y̼ LKkaXFo-F;r3>IoGTE+u+IFh' l$\Mzy`I uPFk!(U29X(PydL3 4Ҿ"e!;I ;<_#V=f}sTxɇwmm#~JV/0dfl2O;AW[cY҈OTSDٺ%|fUUd}A DUZ5^LRZЖjņu/`R"&|"a ZH1'9%ɸS-esafEQ)71Û3>:sꀂjA'twi>aN* 2ha BA_&bE_do&?MSRIo}MP&욮Ux1#r%iy]B]&)BN^Ȁo>4Πu1:^=Lg/_z',:ڼU{Ye6r5? Mfcބ7JJ<'ek:&\{ٸԏz9e'M75nw;>k)]lT6yRb ˭6'DSro}ǜ`;.oOw'r:NP+eR1EA|_n'U}w}u P0vY4a#A29HysQ (>cmS} =nwxp+ m|ŶѴKR#*hʥ'r%?rZ\:P)mʥ?rXo^w\:9Vڷ>/]_g#v֫:) 2f#Q43J&J:l6uM )b鐍և@ wQxQkv3Ij1S%iO0btM2Y@KM-a$EV 17uk#gMݡ5.S؅ `ع.;bYYRr6Ǣq>3Wqσny\|WJ(Hh4QG-Ln)5 ĝK\)ᄭӘG"h@QAsglrD`FPB;ڸ6rKB= q׍ˢp׎V-U[mrq=# ]t߅I?coe,z1c`a逸[lz>\T[yܓŒ1DzpЍ5(KY$d"iY"MsiO&mK'F0& a[vϚ[`NŬ-YJT*-RMA:Qb<3]PYkP@2ih 5d(!t@@UA{yY*y B2.5 s셤d3")!oO{UU6 jzKrY9â9=__gMgZ Z.^qT밠̽O9jd `hf)1H+h%h%H<$Z (A$0EyЇ$ 8CB3e^shlTV%Kp3"ا#L(knbiA; Q խFz%0q?m:k%i'zYvu|+A| ;/Z=1 8ȀJfu @y$"2#6[Sݑk),)Pm&Nj9'oɾkzD'`BќLN2y,'BDgx2YSVMYup)G \%r4Q^p*QY^"x|ڊ`#vG֮VǁX\-J7pTfWłxQ.%+gWwbx`Ď_kN_#_?}GfP3 N_^stʉH!gw>nFTG@0Vh`%H L'j9tNTVL+AFn˨xf<&S㲠eo~/xeq1 iɘ@'ec^EǸ8b` ˙U }M,Sa w~s NͫV~ 2cK'*5~FǨ^\K*rñ5[ 4S10 f1cr9+-7yK&ovachZErŢIR|M5hPGWiU"cDWJzp8`U"cD-CWJ#\ZI2SVIqna.foj W"׽Jί"W k \JJq`Z)-W&e|fȫiuzg~m| ;cB,s):۰= Wi_^|//>Ib!U9! HLZ3ZjkP$qRoV#0)=k g @;g<ۧL'aNfοw='hl^|ܻwfdg{͟9^rSQx,֫uR}hF  z cһ(*0 AhaN6@d0>֕ȿ{wrNq2<1;m2Tz;4y-ŸY?Fy<>9w1n Y&]rZH$wg=ΩGѣQUЕW,8?&j +6&a={_\}GԻK4.a80"Wk 6$ڠR)$h.|E=rZ+BºR}T+Cʅw˅0nǮN10N6ifSxq4¦ǒw7x(6:Hr6`TEd)4 2.UZ2iTB`}@Ǿc'W~ⓒ-$m+i:JG'VZ5}ηN`0\ b7pIrNf;w+&l'tmf[|&*/H߮Y񠉇|`ּSNDpo6vPk^xf%?)i"3:SÔSIm1aÑH Qs,<@֚9jH{3>m&XS0sNU:}'@pK艤da7 lE+R JNA) `׀^-8ų,W6ޘ6ͨmslroH^Fjpn٢9끬h-pcK+62R  ͫ*4-+4oC@wrq3 ~z5Hڑ]6^$`,Szɗ={.۠j1m~#ٻ6#UHŋd`B?mJ#Jk+|!)J%lI\lYtMwWկaܛP=Lt U~ gȧ񻥯lnOQggU؜7y ub/`Y(o٪M=]+.knPEOcN.*')He-].6SF.n3Mݢ6kY^_+JoUd6)C)Z"B((d1, 0Fmm@_Y:%ɘ%&#oJc#[&@0ltil6#gD^.I'g\-^5>`rc[^U%bnϖXuLo]1;i :#9^1;n))h- ȓr]vNv<9H`9h}8Hgm3[tf iڮ]oO-1gBB0x!TC3\ɒPlgH&deހDe]-*TJd 'tt>A H&[_˒=Se6#gbsi5`s.W/*%Gc3?DQ.ˆFу_/9epJgȕRI+ikU^"[pD^kbIc%fCZ{F=<Ǫ_A7 'aW69ka-:e!1N1U2GP`F reM}/nqˏH;noc60J=^/pNE8,>!H-MDNʔa[ei2w0˨_;vcU S1RVxxs&&Po>XGjd-1'1R[FZ'#H'dL> Pj$ZjA_"4"OBhz<.3}Odz?D󜑙4;04;p&>(̳j:ǿ k 915h# F M Wڻ~/87"0_ 91'Ѹ}7̬W3O_T!>~7|jiX2z:#,F-akfL^> YZt||9՘G'Gw f:w:{e.yW)4^|qHNfQrON7C㡆OÎQ:zX~W_8w?:|C~y 88Jao) nc~؈7 ڵǛ 0W-z:ᫌk[qo]i3ϹH7/>o&z_΂:zr3G GyPubPu4#>"RB^t{ԛ>*z):^OپKQ$:,I>l;m$[I,D)-dqI07Q=2DؘP&T٢Hloa&<{$2U&ŧĶ>EsdI*x6*.v9FpwLgF퉵FOo5o-ؙ(/ع5bX y핶vf2S7?^!=/ x H@`Ʌ.@pHVK >%RiF -k3FQ<YkkaZ!dPL?+BJhK^DAT6 vS@̎2Ȣ#Y&"FsTRBt1NjF 76ob>QWݢ7vfSd;,8)"O]|rܡ'BGhTXMLNiz>_=w#軷E^!CAv$*A}W֊7*Z%?۽,}vknO+=t[-QX]' ZJ>C&R$BفҠY]R%Dk{f~s]|GL.s/#jGMƣs2y{߽Y4e|aSirRzu7ۇpt~ 4.{|avZNPlNzFɎyÿ{n26uO^ [7EAzYN&*޳ @,Zz4vn_xِVoyvmy&/ OEm_=]nD0'k8u95?,34jyKw!t_&k^O]#gv_w ؈?QPa+bYTn{F72IxUP&&[|rxKwiԕ*|M%b㟺">]\l}#}KKXع-_|Wzon՝}Z>rE\2ƗNk]'|yхg~-ɘm-.-xO޳x4^{;xOKOf[{IÕO?N ԖIB UJUR\Ԏ ԑw=Y;>ڢR2tfT2gK%FQ#Xshf86HC*ʮ2(DTk،%c;L6BX,|V&/2\tN.'i0z7t~d5 DR|1$œD̎E(K 1%x:ڲŦ6Q4m5} vxSk,sĎ`bnM#VcM^.Nj-րK: X:j&hEE1X1C2MaAcCbjEddȼ%)iM/iv99^GDa]0>i ]6d3Yaq]UO_y\xa'sC+i$ha[ޟn[$-_E.*(TʄjbJ,SѧZMD9EY'Q| EV[M!M^[6d"P Ќp3rV8VM}z|.w֯2kt6Z$Yx0K'P7ѷ{} Av=b F acc %:R 0ɺ! mbLIX jUoA7+P;&i] *:e $"eᔳ%P0KʵjםkKBV.> D,yA,@b1g  v䎽Y@)/Kr崴 ]7avzv u;And>a}@h{׵OnA?T:F_[ۻ9tf|:Zιvƻ;ܣ祖a<˜HWssѼ ,/zѹM5nwKmjSR|S]t[o67R*u^\MkZ\[^庎OXPF(*ڂ=UUZ*HK=GZl|ۅrG ZG Y,qLAaei%?q`X0;e{aaz< sMK6>O y|ucB]5!,bjɧ@d4P}"Ӱ1G" Bmp9F^Gi8Shq G0KzD؃ǃszěi,&, H6$otN1<d/D8=/Sy)M3koFi,hkb\ƜbpBlJ bK4Bd琲E8(m+W~*իy:{_@$O" KEhGZc*uuVٝozoDn։ IR[{&o8CTdw']Lӿ1aU'hXuEn־ʾ<|#6@X4(:1XT20#z{jNZͩQ9jbiuYK, HzYKl<t Ԇxv)j2)gXxYJ/V CQHpsR!F͆ҞPtVWdl^ E bs~zKƟO<'7}B@raBJ'I"Y%csE\t TE}앷{iͦnZnG ]A+Q57Hc: )g|)3k X_8SJEau %wz}䤬ŧ7E>cegvX^ppe ,ˇm΄ =TH%3%+qnZ3 sHS,xe#^:{ՓQdTJF |Q8@1#*S1r1kP\&)Y!4IQec@zt,J8W(*[@"72 I܅Pok6=ɨbHa@H5M7tr?} iwel?zuOh&emp'j9.]μ:ұHUB125c}RA:Iɧ,3Ķ ] ޠxD@)K$C$] HJY gY@+UxjBFަxFlW,Km"$s5TIu\ URRDBhdOR%}1&7Qk iOgD"('% o Yeb]r!F8PÜ=mD ڛlBjk8a,X]oB5 E;4[١߉ڇ[8'GVZ>Lbr|,bBs|hHkN.}]`/Er^aH/5FQYEZvmH%G"ă&7mH뤶VDRaɘUU}8 .&Ԟl(וc^pL$r9b %9Fpnc2 ^h] &Y- W&][eqwh M% | ݄P硽:MBRQ"q Ij+ΖCYީ@5ʃ 6:J]i!Klա%)(<[yA,@!a:gUX=`dЩױ ұzw]]~+xoB[ Իlїӻau*% &aJ'X G:Kci%ou8/i5 \ьZ8`ܻR~&祻lw^6BDԹz!Zk2(3Wuc"mdKi%'5D*Z7nP[ BhΗKH p:yz(sKIpn"d>d EIrRS$WщPzk!Nf;dMܘ4E[I>dlD$NJorBSOԦzx J%l$peCk=W&!Z:#*jg͆uzLDmJo_TcBk((\ZR^xWkM+ :P\,Zhzh_ދ6=3?`&4.< xRxhϯa[,VAxT>*W$j5Zi:N:ant7ہޕdłѾ%j!G:$R*A&+' 1dN"+ t|NOw$j; pM#x7瘶$v%Hfcq[0N1iT"rCNE* ^yDhlo\5KS=Ġu~(H)z}5\ - m="\~ cXh&!jogt>N&ލwDMX͚ׄ^oM$jNagLPak ~RԚQn5{n{zɎNmvP>ͮCjɨ%In{xw3=z^j]*n_漏̏Ge?B;:qwf%Mu}l7MlICφt ݡ ^z͙>Gvq5Zi.\ r{+]|qfQͪUG5[vT8?VLpWz %O̔eLsG+մ֖j]WfkvO7uЭq-|qYLʅL,EI/T{e$J$t6 %I ˕K{%Fhf/5}}ate^sZ?j?{u~k d߬Ǟ% $#"/FV$JgA}zo},Q,(wQ!h~%sc9D,ACP 8<0(*I}6eL[F? 6odV eV/׫J& Aa6􈃚 L_4xQ1XO'3viOC9Ӧ^FFц<:ֱ{ُTui(=4l _BBmSk'=j$RI*.攱7w]j!=#5zX9XRA hL<Re\rwKJK=F)m8dJ%lB[0N0R}!C-GAN9YY/ @ N q 0( wz{dN̩A9bI]! R&5Ot(6Z zPkH uVB!]d!%m-BEh{m3twnfL6엛.`So]o{LV3DiSg&Xׂ!p 7)&J ŧ7b~^} o> E?gj{7_!&{/ę$$xnw`XQΌ7~$%Q)2m0%ͪꧪl?H념GBwvi9HIm{o|_8[Ip+n=,ʙ$WDh*-š(zW_hב* >J ,"X+ӜQD'sZ̥1z,"UZ2l6^q{`smz0wf!ss8)s j\|k'{!MQ8IЎ q{9|?) I0IU]8U!MΘmd {O={PGe<&3]W{0ccõ #bRq"ʤ Ba!\>RHuLT-T_l"3O`j6! AL)^l|)@*YP$tս}}^=s,1.tRcDΌ~xeO!zlc}UlZwQGkJz9p`EYi6hTn \rEQbWcuNS$! SB(-8`G5O''uƉnS*xR %/EUɖL}PbWf|/Iڱ }nBt-^B'o\ȋWr!&?LljPST;}e^ 74x$z{j]Fjj:ǶD6H\@ˈ(f(A5a N>r}Tf ڲ )*?G8_y1-EJd#9rTVM9\+|5;_ՠuՏf8)-jSӫwլՒS~aEfFŋ3F?gF,mÛODxӪI.?YM# Ҿ-heAeO%kn P)jY(jic/d0"3՛Yq: KDjNN:uaw1In2h|wޫ {{n[bUM$\oha;,x0to]8 }7gT z .um$ֱ E 3ft )QB os$8ϻ[X +Y; "Ox*HLe. ']u8RvHT&H䝲 vD%t,/jHי36ͺXf/g9 SC\xWa(C :^ g(i YŃ M=]y)1rU3ď8=B4o_ͧV/8 #̏O{ͅfݗ,9XFb~]Ut&B;Y™"Gd%\:`X,y%v͒:j5cOT;d/Zg`I$aqY*yȘ("kf]t bzB8]J+ fMrZ&j>Jƺ3;7`n~+~􉋄ʽR8g\3Ar%-b~: T>QbOd,nߙ܏Z/{Ś [W'B`:*ႊɀ3't;ujJT0!S&aM0kX | \KP|K@]FqgXnM6^镽_drXRʳR(q۱ke0'n>]n~kh\jr ,S%IrHSL6w(|hgLL񴒧I@6! 7KJ"QE=%[.=?YV(.,VKKx8yc4eAdSŰՙQQX%?0us$;7"dFf@sV߮`8#O0*0!Q}wVݞvc}y n"|Mxtv^,m8GtUR⓳c`Ԏo ŭ=1~wOmݐn4k*' >L~LVd|br28YmWd[-cH+"m:)>N.5}Ul4xi+?n7*F/UcIu9s9_@=ɛ{雷'㯏OO+_`0h%A=E VUW]Cˮb;tI'|~eCn`ԚKcvV n)@~~f.%\/C ڴ/^3vlaE򑬪} BturKu]R(!m@ӣAiRЭ%Qn%d#᧭6RVK 0c5 9A!Byk|kz鹝 KCr8,0YDRh΁"s'_88.-R孕{:TRglgR 9W\׵9?[x4<0smCвW>=*#>!TYX!8ZI ?e/"D-#X6zԠx [0X(TmOĬW`|  D^ wE%K@)*1NꌜBn(C: yKu,n|C[s'YI+ve}c^^o< :eQ}W߹c\gBuQ!*1tny$<]"VգǠ${,(_"8ZgWr(g+c}h7M.lm!zÁ]@4\(O.c)BO*3@cĜE NVF PYʻ0y93.}mRt}<l5# >Am> \>x#WdqI)渇>b4 .,)פg8;! 9HN-dniONщހY\ѨLά4EF]ɹ$zf# 5\.!)R :ʬE(XZ#c3r#c;]6}04po0j 4p{z OaoG矾'FlSOٰD:!*Q A\QH0% U < ؋Y Rmʰc(YȮm;CQ-HE9;/ٛcO^/r-zDE$oR3hLd@lԊO )TXM,da1EdjX%J5C-Im3HyduIQMJ1b@6;~|c_싈1"#">wA!Z mA0 1(*išA- (@'vm(m {/6&Zbف*gR˚NGdKZf_ZװoEvD|U'鲭j|͸d_\ԍqQqq 꺒h/`h$6Q86x=f0!b#.>.qG_Pa[zwkNEW{ ~a]App@FVNJlR{lt|F7Ʃ+&Zs0pUuPukgz WJ%pxo҈WLS`઒ PRkURB2ʻT>'o~yg ܔ7~s1\{ڹLUQ8/#Ǧ*P s-䚂 :~l=!"I$|Jt3>IZѲSF.`Mp˘VTYYQRqVEIr:e)5^p(?jZѸfdͅt0 |G5w u6Mv%;1ϩrZ%C) }OQJE JFty*OvZ*2[-^-={V_nt@ US QS ktNVbܡڮk<+h-dJɃ量)>df3·BӗURbb8.e E-+AgѸvC0ewor`=O>:/z9E?P^tw^^=h^pQ.R=l]>9hxoBU%/qgJNE3hQC>H_kM7 ,PX`R{Ob 6 VEP&鳰(|EƁ1!W^wRDy8a8-ϻ\yt,ḹP]JfӅ%"cK/4MJR;&u伵ƒ9"ƒ B TQSֺㄶE:D s.;3YYa5-##ŜQYf%B -06heMT,wI$ PFJQ5fPN:ԧ0Glۭ+}3eM_P bq0U7SGPo6<ś2ei-,V$B$L2&TjWS W_CmBxƗ|l#|yIwqg 1:%9)6IX> X\-Ipqɟǟ~.jL' H'M}ߕiu3be̛<ܓ.`ryf',N|`EZ6L콥*j0>'&kt&~^zF-]vC|0u#92*-<7GԾ-Wr͓ۃ[=2fW+mR i:#x.$]2zmN16z7 =R_^DTm _! .`Ŗ&]1 lPxJkrʁ](zM@r )ͮ&iЊ7 Y[XuU#oJz]ұX)i<ogw`π|:> aMnnʜDnJaR& IA*ks9&?Lzˊ^6k-{[9v5ۓ N!Hd- v,aeÍEXlQΐ31g?zGW[/Q,}}$SJOf+ LN\eKj%/ izT֞c6}[>V,Tp++>۠T`:sW?Z3ad-NJR~R\o2Ncͯthn^ϻoЄ"ض.%4 lLGc*;c60 RMl)de˺(Y2KF!;Dkc;S+j[VdMQ^e NcJATE\L%Šc0ɺt6c{!Ov8gD n:KyXV#ƒY l-P75T&ƀ"J@2kWP"Ab!qJ=HLN 9^T3j5޳7YEbUtAI{rLmY xJd1̺OSdޕKJ›ȋEV%$#e_ DG(4 TkdlFvdl4f/XX*UՌ7Dv{N8M7OaoG~#IG6,EQjA9CBBe2=!4J Djë.P)lxMm+GBv-l @lA*oَi|Lބ1}i%6Ixj 6`ј"بi RЛlY,:6b԰J k,[Pfb{QbbdlT16#g;vF2c_싈1"#"jAzbBa [e*d4Sà(۵1^ amL :U*Τ7Zy<"[ X4kbMPl؊%Nezv|͸d_\ԍqQqq 꺒h/34P^( YSp1͸/x-5GuT5r7 8pFхew T;itA>6胰T].,d"UVM h@ji&RbF[c4°tȨ#^_?ҴS[)DD %(c/NHU yJN[%kim/Q>Vq,FҮE\sPY,+5%gj;.7?ʘC)6!sLc^ʭIDz|6fKOhqTzg@ 8}#OWWMO7:X2(,zO ^Ƌ [ݙ 'j^Ld9"ZՐ.r3^iM^.qUjN&Ɠ84>_( ǖ |U0$QJjq(X K͊|v;ŋ畇bVɗ)"?uxѽu8tv⧡vZiOp79-VvIue($I^f3%mwfF&uȪ%%e؂JVFeEddF'cB,枨]sh BTWU^]^4fN9_zt;>Wteg#E2+kdҞKAȒKgrLg˦a˧tYV"bP'\ي_ =;]x84,Kged\6gr(XB>>|Q,JbP7tMozr{'ȎBe7^}OOӇw W"z4 KI0&G`#5SSs#֘XO,:6yɼ7Ňb^9=?1q˫~xuXx䍖ͬ ļ (gu5ި-nbN? 1cހvY o MbĽ|$ٖ;"84 D`ց{ MJ#C\i!J&=GvaJzSa@SQeQ&} Az X$ v KMpN鴲W[%0O?>ydCyi΍y< `UA@v/eln~PJ*"õQT5EcKFF)늗-TZBLB˃R2ѣxdW袁I[Fr  4{`",'%%Ƞ?=n'.t>"&F", (b^dTFbqmi5aPH{TyE2%hbꚖeS\a9p}Z˹^ n]Um.ء#)#4iIi/thP긍~*)\<J'|j9MW6?B(Qqݻ8).jE񏥌q[<&֊>a8 fZ2J4rA2"`$07d.C^d|iyiEN/.p\hB4E1m\JwoB\D|2,G9ndFnaF~.@z#ZŖۤl}RoDK Yboph&U8y̎a@gNGǵDvP$Dx */ݛtLNކ 쏤(ghͅ6mk9:ZrG-@7s$GQ9L>'g9 ˜1dmM2PmʶI;϶H$b7 x˙ҒQ NI$d-C BDnw2G26u.xv kq5I:`PQ84p YȲZ Q-ĕ$YY #Jpn%]EI@)?6T.ʐ[I4B=AppA6"%ZN{eFЂS&rET<2Y`IUS|۵[#g}CI CKEUIɔT.ؔ2%!9{fQP3b,#d (Ov yNr[ܲI|~Q`D.cMd?;g2$Af*[V̒df!8Ƚ5aI0`&j +&Z4ĨȻȉ޲FΚrwY%} FVQHHxRz9ZtT@A#6EM7.I*DMz.(5_я葛¤(Q#έgL$PS&h:I[ ?A BLN|$\`{K#>d^HQxG+cHϤr0gg>HN"#1KPF䢗Sk[mCA\,QNڋ mP_G"|Ⱦ aQ9)@~>8z@KΝI6։*5ŕ IģE񘀱rV!hY:{ ^upHY )-Aa HqBBld!UI }qX`Wm/<{}L=x6~}m7QIdl8rGQ+Ɇw|qTօxrQQG,bCԢUzIU/l6Au"?(]D碭^[$ck^ɲu|Mtdhs(U~Yዢq5Zԧ;V/pJeH {$NZȍWRχ^.fߠ;^hqusϣ'FQ8JK5+YZD}ƥDXٕ;D,=4->.iK}ȄW+ocI\O+Jt&R__Zu6F%38C!ƐC.xP!"eq#wnA:79[[#gM+=*/_K{gOV'IοT|0w.=sɼq ڳO\Ih?ʻ/mV:"ϧ.K]=]֣W?㦓1w\8đeív޺t7o5<3r5? _]8>:P:k!9S2Q)b%+e5rvOQ'WYq_-T2?U횈jpկD9gm#l Euq+F\z#3Jk!Q$ o+۠g҂yʑZ;;WSj0b2,#F0%6ׯ*qG:8^_º6%)9T heʬdm UιyV]i'ߊz 62p~ K4ehEOO{aAƶCZ;BZRI~3OPJc ^8R.dYBH)\bav3tp4 ,r"V#g'&$A NH͐Jmm:&A)s]-d1qj.xKMbVh <1&-GZ#g,<|}`-wq+8/b 'PˋzLn9b̄@r =&ƓMyr&9D)๥+>G7+~T&t,*Kmš[c!9t̤h *yiȹ[g 9,-;e8dAZاg#RK)ۏz!P).Rә6,!)R@(]];q_GcѦ9\41TFLjY&2Tiʴ=yy66t:i$J")!gSGC14<ȹ{.433Z]?*m󾖪[mAAL!?sePմ)O-zZ@ զ/ kt!zSJ^ ePdnRT" Y%7DȚnto6dCp/n6NJ\o&UMCIF„nw6s}x+s&C%sl| >Zή\{[pvQXC[w߮\?efv" SԕT)fjbJMba1\@hj /\[4B;(wL,RpJRfe֕\|6gj^PS=L-JKLi4(7&GzF=uRI TɌᥰĕ:VP)O@=@*_+_*_9)-V9\*iP{bIS)4IN <'H]RU+IU7>T=GZ9|+ꠎpO^Kn{rhP 633{;S ~oJAF ~򃳳A?3LP$4dh|W*T1&bhd{rK%5\x@Ze[ qw2G&{QvjW-Uo a Uwz#*o~h,IJgc(s8Z"5׹щH͋PCzP}D1ja*|,{Q׽Z _^ju֣?3Xd8ϑɋ8Y Oglu*\U"G+~}]$Hmshdo1kAm-sz+j7z̰\{i7ɳ ˮłc]oŋՊJi^@uvm" آ&⮇tm»rq){9 7o:!MtMF V7c{,ݏ\wGb~a1z,F;,Ws(&W?| ΋Wo{ޫ"壗q̏3 5&^@ 8~p/i\u>Z=n[I"Kc>v]X7MgއIL`ԅkb^K$/,XTNDUU$dҝ! < }(Ea{ <4yek\0ӲQ@4Ч%Ì)~ߐ63)#<Ԕ4oy-A4F/ck6Y1oH7,|Hcn1T*[ES^^{nD<$hc,soNUF9"sMPZ*"`6`|}MǥU GPy>K[lntgkR|Nݚ϶̱WAVwc{E%ǞzɬMCڰ]ݝ˖,aoh+ƁZ "]|]¯F?NZ{6޲gPiЏh33B.Ṅn1wrofĻ*ʰsֲ~ e!ӥM^_OO㻼;xl)Fs[}*KB1K㍿-Vl٫Sg=7/w=`E?U_w{q֒ԬvP l37])A7jkl%-NYwyh|ʠ n 58(EH}2VH,2@$c{<߮J=Ý!]ɚd HWÔVUs,7\x/8jt[!3'mE(֑0Fq]"85X^ToJL*%/|!c_hy.hޖ/v_|ћ=emXdſ}V;Su_O]lh8=ƍm/2:7i g᥂Y$|8x}%9\AGu8%w~.&$gKSB0$$cJϚkfDU`CL )@.2Kə (sNz6aߪeEGEM/ӿx9↣ #ƂV< dd̉4.d<CvE F12*3 Hd]D+^Ћ#B &1B*i-pT 0+3j5p63*mSJ`mi|vo늎|:WلkQ+yP-^SknCPmq֕{<ݖhrѯz>0R H|/%Tܑ/dډ( -JzV& 8Q|0YF!xDOUɨ)XqbjkSpmʤHDZ+TmX ͒RMV]e, Q>,CTp͸rhaǟquuw/4cI% z2,#r 僌Irt! " bRJ0Aeܩ(bݒ+Vfe(ƞOElJmH9 )o&ێY8s̢KgףdbԮF;JmJxG}`{(^*XeH2`1z^ 6^LL&(@ 6>4E!(д ^(l$uI/|HY%ɨ-,[~CcW+KD~GǭHISs&SYe"Q L 42cD;zk]U%1mlcT p \LH4)in$ɒ@i8TY"VgDxՁ⸙2!:QɮrQV(r1gZW\W<+ LZEP3C+I$љcǣ\|x(uOH^kf[U.p|)Zp 8ϴnxޱޟQWwOEՋ97#<4+yv:9]F iԧq[jK4l q0dsn蜸(p 5+{o a4}{н^nc[ˊC`\1x\OQVV.9Gowik"^g`g&-yLt7蟦veRyK,ptʧ뮚b9ůGg =^9NAY{|}ʙq&ϟowc,f8'|p#sȠhn{<Û "Iy˯o߾:_~՛\V`ҕi =x0_miS;Qo3rϼ Joa-InݧWpKY8D{A[gR&:iq:D ՟V/νxՕ2]d.GE{v4!d_Ka{>K$L xQIf#HIHIEm\} 19Fe 2HB0Vtqrws(hbx[f ʽ+H5j1DJ+"^ݑ,afjA ^/:W;|A<`3&u>^OW +`\7s8SpD.Eq-~;ޤ%g8SlNWh >bI$WmjZ B#8֘\u!ɥ@Bvh襽1_H'R&d1@p4cAIϊvP8YfҖ^!Y2qϳ hcp.ifLb:9!jgfpU P:nZUzFp-zf=M5pdjmx*ȮǣVB-5Ϧp-iNG-i-`s  0W3ƃn^Sr8fCβ1Az$ˢPo뻽catq*_֨ Qn^臓{JcZ͝q؆nϫ9MmE_isO✤lP˶m3}U8 |C7/b^M&hO.>^\,3HMiӀָxUk g3+y$DILuc_5߮V29Ve' vL9Tb>ۜ@ ${09Ec(^ǚ`pIb-J. ƜF Qf^Ɂ;VgKU+VWM0O+-3~1‡=*OT']]AcЌ\>'FuQ0bScN`9\f h:#VY4G>;Mi]LA5ƫg):28rk)9E#Nu9QgK#Ўq"dL0A8aǤtDb!P2de>Ζ|vpE߯#Jt8F4체][sDZ+(KN}TqR҉]\sh iUY,I\HnԪ$Rf==_Ӕ Z/3D?(NpĦH%Uv"pRQ[?. t#7/NQ ʕHC=bɱ_196B<4u6JgA{g0ٸFTBީ`$ t^~'c*?t㐧J),Kt^MƳZUgi"RCfKW]Q?H>Oa6i*~U|%0 ePYPↈʢd +/9a# GT vwLu$DYIR\RH4, ^K/]:ހHN;kd4 6"$o5EpI{SE}#EJN-Ff ~g9AiUJⳓ.>v]|6,`v]y=d0Z>~6 &}IK$9EݠOW.q!XwYx֣Am>JNQ9 `AaD̉\(5Hŕ,d=cY1$SI8ē\xJ<80Zk(*Kc{4 3Z5r}@ =Od}ACALPSDq03ML!μ4B;P/{:~o (der%JL º~o_`7xW?^ZI{y5A7AIhFpbh^&M=OWfOqbU <MѣBֿQ1U#Ļ ߑKo vvp_}M,=9X9>]Gh w6M._%/{~IΏĹ;C3{(~|!`;!+ws72s ȥW Xh9W%i QW^kjE4bg:9 ҝtwoW%&?[iebwE3ydifIex_X@!~YQY-qV$vM̾Je}4{+錈T*.twV >ŨIN <'S04O][HZݞkr. g(ApǓsRÛ`}}p嵾 %`N';l`  hSb{ˢԻsb{Q'_4Ҋ"Vf1GvqiAh#<*% PGJ &wkg#oqD?۴8ZЙ00a8a8_ ]?#9-P1@kYq\A\t;fGȓ.칝Nh~''uʻ9DD]Ih-N9I:DF(AA8ڀ&.J՛1s_3|gut"DВK,Rq!F8z`Z):yS b؟bP4P]jG*0zI1Ŕ %'JI"]o".#@H5竻Gz=A$?nCj݁#O$(whouxd LRͭH0֫\Z0%k(p&ν)xЁ Rg3k:SDdn}?; 89N^eT(FEFB1hoZjc2h)bRX~.C2-(2Dԣ@Qim&" J< p3A5Oq8'>VA5Ō(W$b X@'!SWr}J'~؃Wdӯ{8$l.԰7˲gջ?{GQF}y"bo)b3ޏ%%R+ |Ahk9.˧6~,Oc\O_>*Aw>tn}gMןs&6hzrmͿ ӛ1Bݧѻ|xLrԿˋ͂-ɛ%vo!YբuQUfnyqxk~.jzkN@)o,B\sWu #/|x:M<u&'o=xF"EW{_5μ_w^"5 F=.e#A1 p(aP0G ~꽛Gz8iΞ̂nc5q+%ŗ6?blKmV [m`gEdaJ= ^MvClcbSE]&`%dbLhg<T8=S4/K|+^.s4ąGNr.@8!@derdej7T25qD56Eb2ߜ[U:Y!HD0@" tzWw͊&ZAzʇ| et`<\2婌>I"&Q X"7R@k":Qq[.C[PZn Y*l#gZ/hnYֳsmr;1۷;ʟPӱ㵱|^0kmәׇ1q\y-OƼFrAS1ZFkwc^ Nw4?zsw7z_~0# B9gϯPlOv9~FxJB %8+=AwA :IB) 0ZH(}mE"|9ܚ˞/fṮDy4*h 7VhCלv]ΰWo>fְGdB}gS+ZO5T7a^2M<>q|sGWPH^xm$ArJ&t#PԞ+q9mRlc &Yo&U^Wgk_}k?hL`ʘ(Һi.f_iw& {$`ž/3y7ӵg*fgUkI(x:-5ٌNYnla*(h8Lku&ÌzpaӃhǤVgӘ)O"e@Wh߿?-[7o=:.h[+/&x[y7Qb3C>57w[;ŻW ڛ>c໋oo f U]s?Mdu>Ƒty-#hi'#JV _dh5L0Dd3MO{'XOAϮ<:9:`s| ë.79^\cҬriI!}v?abF2Lc_ܟ&';O!?P/?K߿;q'pNȗ  Ix-?}ҾҼriKr׋d]r%)^=Յ&?]|n?aMZ&Z~n3T|#O+ 6t4b{e3z䞅x: +q KW6Eq۲gH˖6o\GdX)D,ֱTs3.ee((Q:s,矻dpX2Lle\D53CP*D\R]&sJi٩Lf8U&m397=ųq+`M׸5b2A ߾K,/tSK8"1 ]"d Tc ҚHh*h䝵ϳY6Qɒ0G^B'd%ڒc tUJRmd E&e eщ%>9@|j,'5g<; i:4QVKGԁܫ5=tF|>wY<)x{hקffi&"D,:יRbI,R1ؘFM˔ ~K,,(e"D %P*AA!Q6 UPn !ZPN&%V٦ G,O^)cL)HcH"rHJDcd] ̓]0AK!m-jeqXĚWd B-df66gjs4USQgLQD2) {#m[[8Yj;QӏW O^{UV=ǧG_96XyV,Ed@BBEʞEƀ4P"֩^P))*ֱpɮ%E@&؂T:ߘc7qN3Zaǖ\f\{#V$Xk1HѦD2 Xؔ."C"Q $&fb}Q⒢ZCAĔ dT1?lWꗝ#+c ?刲1G#G9C%[JJTX2U JCF\b6ԶBXgX7&0ZbU*I5LkV:M5ie4kǕۺI+p爗į:1_gR%[g3,ٖ/|Q|qa^5]Pv_;1k$(+x=f0!bđ/~ +v ?/VQa]ftߑՈy7+n<c7_(XZ՗i6r][L?y#NfLI3$UF ^H#nׂ۝ `&ڏPR6Te7d0O<~ݢm+(!G# JyI].$j/VQ2$q(f,̢e% ( ;YPBKv" vL($MPD,BoU}dݡkRJ`.%cP |XtNB۔1"f䢟k$Ի6=$j=_"JI%\DYx]E9Arkm&+.`M1j#Ds^e gј` hUilh慻Ghn5a5֪vzch|scyk_9,WZˍ }@_KƛNBڻ(>1FCybOl"պU@Xa'B|z9bYWˑt[KwIm/YOZSHLG"XK`ii/lyCwHD,H^KeO^M_/R,E/hJdŒ4ȄP &SоޚeF)v_td?)Ʀ(#P3.[3:lFFmJgQ!yUSka8QAyr.mh|eOOE 15gC:۹C D;И%aM E{*ERD:҉,j~H"rO$X?@KkހO0 3%ڟI˚<88ma/M EEfzҨ'=)UeG5?KjmEdaRɬ $Sց5D<2&c:5É߰pS(1ZTti6%AG K"2 1:#jg{kFjkgajXޙEPd׈o*E)jB_*Z)8U%RpH!I>E<Jۭ↭0d[#|FЋYX1-S. QO c (`p%h\.J:tY|gkZcoľؽ9CA5~'J+Dz V@Rm=9 6)Ug1 X^ nJ7~(xg[~-Y&,G^›Oݮ^56;ԋX~=b2ph~ut7|#}X&8QQ[乘si7 #os/nܮvazst9H`^Og\LNWēƾ&\g^KO6oYzl 6t nCQ ZlkC[>U:ʎ{rH˅_)& M.i:9qUdwM&͇sA]Mdͯ5Ňn~ب7_Fơˑ;г;>>k\t_69oߚW/Q1|!I6!S¨L[,dE!DgRQW^w1h {1[SwtSj9ŢkM,6[|.Rr-"#:$DPt CEKpŢ ([ VE(zMgaQЩ{"v7dz@vnRq7Ǻŋguy|Ւ)E N,+Ō{ڽ\Zv֒8&KaVLIu* T#M)ɺ(h |a1ׇ+1oN8NMh:k2xܫYb֤o[DD߆I֣96=3|NhPv.켨J&aOL#钥uM,kPYҊ: 5&_>~:UȱZʄ2 & ¼&BPib w%JĨl?vf nr04\k.A)btZ.&J ;e|Đ?X!\JJq)3֦J +2Uy!/N_;a4* դwP+ŵuVZASR>XI \rE9X!a][7uIN~[]D&&ͮ[_}j"N1B V+LUh (L0qcl+H#xa(aZb81H W63(g-N)dB`ICR $m.@%}m1s8`ɻjop{;)+iV92c|-(f$uH`mg#%[5"Q`Fg1Ӱ%jg0|zZyzrw$J-wHbc9%ˤe-TU zXFKAx%8/e)?@Uї]}PbWf|i]`UʺhZM* h =j+jϮkȊ*H<*pq&t_>UgHcf !NK:  [zw/d\o|6Սx2.>}v'5!uOݾ}ͻvsק H';5ģܼpWobfQ&b: ɻOở -U67޷9_o{a| #fm>3WgW[*L hp;,}4PHO .Im79D) b>^W /&1,;e&W!KО=~=oF`2OEsbMR7~9%0AMEi9=y7 /<3Itdv S)C|?[U}&/,K0'œ랥lA939?|I9\#3`0|fYB?eY0 ]χ!פwQFk]NA^p˅m.ist&Qs  SyAUP#5~3-ZS1Q㼉:ok4 ƯZv #D62kd}9r /u,HSÔSIm1aÑH Qq.kE^FTE˂6~e @q(.s1y"iuXgg"+R,w ;E}, ]Kz^@]0~+]mCB(mGPW74m(^]X{jkN';OrQUK[~ ň6XQRrǪhTU"M'oTL+Y"fɔSKNu(OY[05+0w͊bwWjVT;b',pQF$aN'㠲T:"2nV(Cd"#ɍvj=r=V)n[E-ja[aFhNPf{dTO`of8,KY^1t'ӌ׭Jt U؋̴Zوi*kJ:*. N^oWؽ 4v(^-kҺ׻=JѻOc>6d:(D9ͭUD95O{4;0 ~${" &K=MBqO1W]QaJ™wwo tuSh:=tyϐ aU[:t Yw"o 20n]ER4"\͚BWdkۡduut8r#`CWW"h ;]E tut9GKWXpUtJ>vQҕJh R4"\՘P;ULԝ"R+)RM `ݠWɦЕRw(YҕBj>#Pnps3G! XL7_ FRI&:{,cMr?P)3_B2~ILgwa~8w-)ۢ-ZUV(Z=FZ S&C./'FL:9%'+wmo) 2m"#,Ue\PSL6%Qis0χv弌aOƦL(k#1K2{lMwu0tB`VG_&b#urKcky%Yڑbdw(,0 NfYҩo,^+ڋx)T8?/\0׎Y: {^i+{I&\L7Zg8E)AUZ8d%Z @RFO,uJÃ&ݹH-I\XƸn huw#J*[ ]l7 G7"\&BWuRΐR4juj5n+Z: ] 70;yrhU Pr[:Cy0 4"\BW=]Etut5"`ٜU0 BWVv(VЕFHנ|61"͉]EL7"Z)NWnW] ub\>v`N]db;ĵC{bj+̺b[kjסT O2\=]E tutEO+l1Wm@ݼcHXc.l KZHY:$e3d>ƒHUth;]EuQ/Q s*CWC{rCZLXV01tpM(%k JyX[%hEW"J[:CDѤP;9 muRVw(9i JIjs;ynl2UHppV]%csS$VZ8ύI…z<7֞Vx.sHj|vphѪ/32L,TTDW6hg@ۜ$.@ː;]Etdo8nXVp:qUОPt[u1CtXΦ" !W^q޵-7r#_QL죫K⒊ݘٍ8ޗ)&)'(^DI%UU":(d%xՀ&tlLoh \KM ā"M[)wRW~S)C5FcΦRg.]-Pj8'5SJO0jd`}9YQϢ3aFM:pY ;pK#coBW] Jk J]`d7tU *h:] t O 鏺*p7tEh奶#oztVO_~fp ]Z: J9[+b?sW|_A9]I2Z٧`[{CW.- u*(7IWVk-QWn$"*h/uPq#V ]FaeK`{3wU](fJ#cf\.jS3B;G6C;t]=9ՎT#oUJc#r 6M`74]*.h/u@s# M ]8>`yo/tUЂ:]tJZ'uEaךNW% -X͹]uEp}Vv J%ztƻ<V.\fh:]]KUJ#HO0XžUA;?^P!|teЀ=bo*BWd]rgA`CW r]`! Лt>ߌPJ]}3tez1.LW fh ]5Ci;L2]=9 UAtUP;[+`j'XX6$̍wUdJlCgh#;}9s+1zga+.n%69Qì}9Nn?I* =ۣڙsz??fN*ڕ(q|~_ Jhwu|:yjs;2nnO|=E]KGw7Q(ђȏX/pWJ8=;-J߿+j~ioXSJX_܏m>$쇓o-4OJižqŪ 3jvFd 3Տ#X뫠zfMHNXWv]`Vjzs} ] ࠲C & 40AF$.p'c9#VT& ezfc9E[' ,ctY`Fu[jgP^N1/{M܎mbI4:GnolfIX 'lcJe,/x}ɄF /\_{][nr@.HFZVbʽ5E:R}h[O@2!a1$V智]ir H LQ/10T dRuU_aړǖ w) (PxgRc qy"+yWHS)|1Ԝ1=hh+6:I}#6'<&|BBc JS0o,*C"5֓g8' f| ksE.ȍtn|_ՆQqNF 8T݄)Ǣ!# Ũepw:ci>(K*.-QھeǤG)*wunI?=]tq]}=+ڮPz"cvWO 'np?xs" n<.N?\Ս|]+(rn 8TjTqsUt\D]qMG/Eh9^FN8ʟvo~Sv/]QwKԲ觴zpndyUOM(qjQJO~o/*1_>K% ]4"! f v}V/9). fUʁRZ؅%l%%ut O*"1"HI`؈)" "(4KݹVMBԮfm2ODF$@NɁ~\k[pW v(]KQ_F wğ qV㑯AѼT9x͘d`ډlɦ<5ݤ{~{LY$;H֤VX8i^qMq4=5] &S ){DOR3GbR ;M%܈sA!tRxG0wE`tEժqNOT-$ɧYWH%+*YjO=?XיgԼS=̧m>==NyczOT,n:G$?2Z8?ODй5M7E nWKe28"yZ.7ì7 ɅTNn6TZ,_GkaDkv(qn65 6VӇɎ7_ӏ,VFus ѻ5F` R<`rCF5JCuwsZ'gai?z2E$|4bP^ l&*=< j.@xdIp42,1,A Bg.P "9+sa{kpΌޅj6CxAne>y?L~h?]k`БlYqM_hl=fʂ¨O%oѪsY#Yl V)9HRhs))hى[oUz_m|g.ޱ))<>;R'GBCUvStȔt?f#!vPjlV_g_^TGÜzG@ж~Rksʖ$J:m*l=r~`r 9{礵FaEqdzɀ$e2ꈊ$\Ly cX^6P>v6!@BdkBdэ>ަՒʕ#LuݻoCyHW}ɑܽC5!+Ocv>ؔ|V{Rq!qRQm jG:ꈜ:0֚ꬡtTTwyےjh,ƶܘ Ŧ0'u%8+HXyY< ^"vfџ5hr- ܨ{&| YxFS2peu,B*s rr҄e7ct)#f:C&^VA=Gkm{CцYkp ̷6dëY)! z޷ٷCVBA\FHvgٶ^\Z=?[A)+4jTi[ArRJs& +zү%k j"(eCl ")C $FJͥLH hJ*!q⼬nuIf鷷Yْi ߽Y-sx2m A*"YmsYU▐1 2\]owzݚ:Ȃn[dtTb&y!P|õ3Nhoc!m5, O.6a= ~=[Olu)e@΂+m΂+p,8BYp%Ypo,bm?΅[z멒ԟ/79e LZs9m9RLc,*+JGa-aanߕe9N)SBƬ*L,!2 //?{ /^˿a ͮ69w rU^C |x_oiPYYKYAz@ K" tdS!0Dڷ{A@ޡNȹFt% &TY' 1FKGNk{_<u7e|lgaSG)VI+ MI,me,)۫!Uxڢ *$ *x2J9-=v5z.zHm+9)bD"rPGCpe38 $m6`qxk=F"Nl/cy"#}0Ƴ#r5>%F!c[=ICQPc0)]S]:<s~=| l]m{DoHnoRq18AG)=g\[sGFP kn>zz=eBGKWTʮYQSqy:mtձ \5/lI,$2-xm\v{L,c0GM,+&9NxwnǷKw.:%Vڽ D*$J*P#1. .%8Jks|xe>/C^-w¢8Z*teƠȕ:+ T!Jb-{VJcu{mx.mlX']^j[gvUת=#`TE\M0q0 J%f.ʌ|/"/_7c^0<`l4(g-S r U 0J%2>HP'n:p3h$s^M5? @|bóq;EblV>%8 !q2#%[5(C cʟ<xrU%OR9<9# yK@X$`NEi;j^tUZJ!8Xr|>J)P^8) UIT}PbWf~w6oТeή? EԽ@Qך]M \DvilN]\3BLvl Ni-A^Y-B I⼥aQ{1~1ޖ`QrrnR;!iCMe+۴f.?ݘ=;N.aލƅ'Ynqi]i޼I7p]ΰԹVޯa\: 'ϾR6yY]1< U@:/zvosy0٭kLs_?d&˜~nRe.G?Jar~N#d'K>K`fxrd.{ :h!@3hm3tITWCZNSc?¡SX^0# pIfəu#gF,׳OsS0kOGҾ|w^e%ʞJ\=1&97`J0eeXf`4~N&( DL2S}5a)EtU ~)riSm.D jKW[U h달5"ƘaTduOi;n;шm){#Z4v$a(qT:"2E;*1QD"yx@w B\ +7‭iU"bK(,m=3{ Jf/hvF:\;M.G T[xe#:O[rwXe~l`8yW;3>kf1{4x!XV.̫ڇGXЁ+s 5$Wv)A =ûYmɹO!>bUp+ES:=Sž@H&SDF S ˃y}VÀPf\:Vr%ӔW0n=wZ@g`X|ۡgFsi:WfEk{+&\ʩJMEJbwֹ:\ 1Y4G!~l=u0>èdOWgɴ?Kկՙy}x$F'wF̥~Y#`̗ 7~.}.{ Kkyȶ^[#؜d )}~]jzZl$4Ά٨+UFatW||})&ӓ/`$8XZIP&ѵk*Fךtӯ[>}0Bj2s@~M%O֫gH-5 i V70/KM9r%U.})B6 Ax`AG_,c{HxYb7#mā3V/$z7 i=aNj#=yy9{KF sA$@F4=w)0apE qi h*oT|ɞN9w9/O?)~sCvvvcgs-j4n;Kirlks~6Pix`,(3xn23ŃoTZ`ԂCF.Gt /?YN;CpN"^DZF6 mF#˩A %@M0h(TmOĬW|  D^ wE%K@)*1Nꌜ+6iv>H҅\eњ1,=/cgܭ`-Vbm] QD4 i.%$ kpñOo wYYrAYY峜.y=ymiwr8zmC`)tFxèz62FsD&)s`iG(Kd68<+׻%rf8%zvTag?` :@Zh_Xl0nxJYy6,?4pn2X JPKseCKFV5~TПM͋jf43E%|J4b/kg)2j8H.s _zC~htr)([п:?⬋@w384;-:Cʰ/+ x{neMJ wI]- $*8B!ZJ3`NLj9+j$UrJt*o3h׏a|}ؐ%3(|K3uJ{A- \>ݶW`ybNծerXpr#zؾ&=B}Ӵ/A-dn6f,`XEt~>fʘVݧa_A /3V,Ph') J# 'eb48vham'Y R;,\d1ߡbb SD1҆5'( EO0,h/Hl  mYl AHRVa S띱Vc&ye@Nͭ:1茜5]V]#6wH\QR1IL.z@8 ̈́F  a)uvϑ!I LQ.(qCvY$PNbXXH"VSY%7+أ]O$tnq%,8''MYlLZ4ݒ.1ND 5BGS Qs' 1 oel3T2 %Z,B))LK9礏oW12#E ƜBk9wKJg3Te­ZșL2nɂ],̀wGWan>|d͓<(.`j"ʤ6<࠭#;d !#11#f(FÐ= ֦j 6`^&7n.DXbwF~Aܙ4]qdžR{1qR{Ӵ{ XVqEH* ,٤ @a&=2j: iPr@0pp4HI \"!P@=rTiL3rRm9_,jʺ""j67ZX @3pqcS\NWeiTNt$)|ku4|[z ZSt4֮* ]=wyN Vngob[P^6=Y޻ z{vZەr'Μ(N՛?z҆<ētBtbKC"_p݉K_JY M3\f_|9M!>BNDWZ ]1\BW6<]eܓBWCWЋVNa1tN5tt(m:B.yR +bưPrNWr.Е:ڴ bڸbA:]1({̒M+xt((BWGDWG @W)ZZΝAsgڃ_jgNQO} w]m87'pkvjӳ3b0cNt-Y/3y}%w_kxæ zaw?.r;M:w_K ozyj~']?⛣w:;s狧N(w{{? =>nn{U+Bf>Ɏ0>SUqdXDq~ưLayQۍ~zgϳR>0_p 'k#yWݿoVI=kGiQӅ|IU:z7|+ecnzIw >.>Bqyۏy5.uq>=o6{=2U ߒq63hx(+g,Rt~g}l%$]rQ-@BBmNKi6 T{l=O0w6HGР`5P \'J-KoC ⚳ŀI7k9q^gbhIاrx9$T,\mESD*ئL()׃M]reGCz%LsƔhes\Ө%gC hB0dG}woL- M*Z)*݃6msvXmM!:ΚJ*0B?|vޔas#Qk*LV<D@#.dIo_\ǗbJG@֩ X2|,Ŧ!jIՆܽOIYU)^xNua0)=Hn$*Y+1/Ш#Ú nN'"N܍ͱ'Wsqu~1Z#MƄڄܜ`/DU3)T- 15?}i"%$Xy2jj5Ѫrʛk5*M40J cPG)$XudI(!Q!s'# 閴P4BZ%36:aL9'X jS HR;WZa7}ZqDmR&Jf[zsPyUo< y1Tpl$_X$FslY6s[9\baN7/t\SWH% ݑU)*ޓBd#^i) d>A"J}WiZ24Hc9uF1K&B@F37=y?]8B@B͢6CA,ȸCMAC&뭋@44'X!AYPѮi[V |u肾.4Y.><[..Kuq@8k>t klVaY{޼Е=5p[O8[}h41k7mLC3ԴSn/ɸ:8nZ̶@k\:t' :J'd2'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q&@9xM(u'1:<'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qx@>BAv9N v1N u@@I@lǔH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 tN~*. A- Q- )(zq(}X%N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8>}9xśs~Ԕzx}~@uUx~| Zq I5n\b.qKw>M۩D W?xqAVA]eo Mu#~^Р4~#NV[zY\{AfQktNQt*&~}zwѸ*ͯ[7rt|}Wj3 d"v/׫?$o@ۗ.zB^zw%/t/$>k$wN1o7?Dы&Zf =~ۼ^/62kz=Q7#Z}^;YLN~vlQ!/Woc/^g~F%ǹv'ʴ]ݛg7i,3%B$86pb 1hHTKRQr6|sz`ug4[z ]= eEW t; b.nZ ]v_NWXc+I}H{#=иYp](icm-^V;&ּ"P/rUnl%{ {{'a!\ s+].}^`y=X27vB޶:ƅᄍMOX&g*&y9|RB){T穲,ȇMwrS|+ԗVb<&̼*+<^qT! S!8'瞳I 泱{(tVR3ƵluE{WII2K =}u---giA8#+vIMEWHkEJxUuҕ6WZҥ}ԕXEHWl#+'H\Uue !]\ 3D\KFW@+Y+i O#+VΦcĕVG+i_uqYCHWdtgXj;d캪IW=ԕBZAHWLIWLDvN$]]UK6dDZWanuY@>]R62H^W&]R (Hf#]!HZX EFӈkM#7k(H\HBȹg8TB܎5 я*һJI0k! q+uչf* h]GtQ K+tuhs)Y={:\8]!- )%K꣮4o<)*و9#-R4>{hq͑k)LiITIKFW]!5 )Hꡮ"xRaNz2+Ů+IW}ԕVFݱMdq+Ʈ+&]PWF3!!]d*BZ-*骇Zzkj,z2+Ǯ+&骇rIKi QTB\CFWH룟 IW}ԕ7J+>:m i HJ,z9dF[U$ \F0Zd0ҳ֮ddաEϕ4-v7The4n Hq&5 `'<M#n盅Zҋ^jImyět)8Yel:V=5w1C Jo{,8OȄ )YT1 *N.S<)3 )Mq QSl>)6RO*M2Ŗ- tBQj褫 f[ZB`-]!"shmBʤ^  IHWm1+-]!~+RtC] 5dt^*Bڮ6PZ&誏IMi2JI2B\ iyu=ҕtu]9/-ӄtR;:If i]WH钮+ !) et;*BZ#]H^TˢWCOÀ]Giኮъ֮(edRRIWRx@lHɑCB-QXw~G4 WI*FZjZréqA)2\]!mW0J#z+ťS3:`VuF%]PW[a%!]tFW]!~t.QWF08!]!+dFW@˹]WHn1RWV" ɴ'+ĵ3 #!+W**BZ+cPT+/>+:B\Kf#ڑRѕnYzduCc]᪎'aC JJ']Z\Z!]Њ *BZcPJ+WǓRGß^!xp7@dE!"(1lf[<_'L4LG!"JΨH JbH jsqsjs=5ɬ`b"; OE(΍!BH2K +,- f/- iiK RIg) k]I^WHVB{+,3n0t3#Tt:HieUu5ӚF>3r2bVu:+< B`]!1Tt.m1@G$]GWV;&hA~BI%]EW*{NFW{1Lꏮ=X]!HMEWHוѕiYfȌ`S]ZWa`42IW=WP1hG8VSc?HW@[FFӈ%M*vM#rI=ԴZsMHW%+ayb8JW@|?R]&骇o 3 F\c i6骇RNMIWl +d&H "N>J; +ZGEWH룟 aIW}ԕ)JwD+:kWAE+>骇^[ ; h )Oꡮt q+<R6']GWZX^BX|gMj`~b9iy_d_6/RdF[U9ΌLs\*Um_ |Uwd#mjZgϗr ??-V|)bT٫nbUy0w[yY湯FQ뷿ķw7W+jZ@2r{?6^g3f^mD?/ֱVtlpݺ\\q02z7P8c RnԿVHz54u#o-Sڶ*smmUwݛ_ F;ܪzWc|t6>r4(/`9vO%vm}j6o̦`1JZw7]!7l+7ݽrSj&[tWOkUPߴ9b6uᝫ;Q'8AޚUc v 8yv?$/iGIưgqp&- y_`dz'n>yѷS~Zʏ5h.uV9HO4'Jo#G`Bp҄7n${a2_zWMx[w Y>Ow SCKs+Yo>,bO9ܛSd'vRcRۅ=N^8>qz{_p#9;o;68Q磍aG1 ;\}O޶}9z`m7*rn-.svU6HiYÝ\Yшݑ`&!FBNef3K+넲V/G|͸쫅qB٧[ mW~()h: _3~˶r8$e3"^>/+׃/K+7^h;930ݮc*l<0R==Qn1H}{W.\:pݦIMM1[b B~B+V+GHk8L_owtq]VHXQMr^I_8WnMUzXU*E9p/&&?g Fxt9pV~,qyάwh`jf|]|]ϙJ?_DWes^~us5#_׏˫j;â܏wXUeIC|lUyj"7:"a.g2թw`j-uPl //S,bFzt^ڼ0ұBb'ęYXa+&=֌܈Yvah Wy 籶.z5{]o[GWӓ' 0 bLZ߷%ղgÇfUW]]rJkM9g@pzOWZy{L ͘n(,8+2׿q=Glr]VĜ)wb\._|8\ef0*8]Nd2%5Oqp Dq(q%;yGI.㉟N"nO32'8rV3@r G0x@pt ?n6oo x~ǎ7gObANҷfm~m-Z7[wv:<]F iԧѴ Gۈ\`aENXq®i8Qm0%7{+Z̛n~l]T}WɸBxW<;?Ysd0.F| юp0e$=8#]lFlaE5lOy'AO^ g;g7*Q6ꖵr,I.5 H|:j]S,98'|PVvO;5|GdΆۃ8<'q3?ˏ?__?x zi3=0dsȸY44 cp.ifLb:9!R i]u Rg}Pqc4԰%ʽeo_|3b[Hqzq <뇉">7]Lp>\:(Vws 9`z=Anett̵'2( (Im/-:B=Ypv>BUl-\қg[;B>zÁ5 G F9QreL .AB霹,ǸV9kdN[!c4ƫ 4_qd,Z+pPJ¨8/V'$l'OCnm}vW?n9~?|oqP)~*q 㝣(?{q۲^˖GзC=Y9$f}<ͻzۃH碲-:;@^iknZPoY'i-B~3R qhL9& \>9nb{ơs7UnJ7$yrrJXIPH !\21j% g͵U}#je6XS.C>Rr&s/Nm<7Xm:24.(Wׅ6J+Qu56Gx"%_˘S4+d 4 uF?=,BXBnM<%h).;jU=,{i1NLBɅa hcRGڢ-;[DC|@ɼtԳΕRiNv*ݙhdFk#9tp87,1JlhX6dŕ TE;XnDʈ9E6"DU3I^n/G d98.HjUZq6yRϓ>)YI b mZJCMBbV5J*83܈S'D蒧Ey"ȷYuﲯP DEa<3ɒ,%ѢAY[bٴa;Sl^$O?AjWRd|^ϛyYl{^‡w]/$wNm:eͥ/mwץvzBӠ?1J>4nuҺ==tO6v>ͶCjPe*ûx?d倞ZdGww;|Cz.:qDk#[r˦6w~sfzBL_Uvr%J NTiL|!a?菿[ւņPoz3B2OI&)M{~si J]7d1QY$ ?qNt2VI'ԖkF)[7dK,[RLcbY`q.PQwژE Y.Rk:9Y枅X:v iS^^= z^>SC`_'~r3^#zZrph2 #F(]r&qxpe=}*xKƻУ3УR3jd΂"NV5mN6lb&)W`YXMHLT2ʟ6$Y&CW̅ƒ@Otd\eZm:]˫OKG..-;7|)D~<ܘH`)+IW# sK1A'ff%VrQU,uL fx4\2;)ZyA: \(E}"MgP"_rE;0en{ў͖ӏM7Uff9KKT-=w}!3]kme/ۼرyȇ4I/@-o`jk-K$7q{8]eydK Ԋ43!<|9/wyVo`R˲Ns)M0`x8WX`Z[FH(򥖜N)bx$8_7F3 OJ`hvS3({z[X =_.>6'L onn-ѐ_v&g@2`k U c,UA#A2rVESJx@UH6A$2L `m~<HW ݆"DAEŧ+p_Z!?J>;*姡+=ϮM~n~]z lI;dy\[@n8YDMywn.TIm36P lԚetd)͐ bӝA)qg0NmT. ئ-(O9-J-Aƒ%:+LgEʓMEr?e;rkKaEzg{!s1%LhH9GXaYj}ؒWIw6l}G}Ar_@PQX5J*Y0R`D$+h3z OS d (D XyNXȜ<$YᠨPD(=:y]BЉmAn̔JmWO dF/t[dPUpA3`:S'7&r#$"2#ۗ0wnMcӰxm~f .9yl<}-x@`E%q<^$-WK(}d3a\~`m:RjKf(52aB}u{|=k i]9's?+FXb g2 F+.`Ԯy5Z[<,*T`FSiD)Vf 4{@:tև@ wQxQkv3^qwNR4&' IɿaZͮ[mq&eHKs=ܳQ/?]${B$4(q&)~5`%syp¶!41E 62(C/*GT KI@j`RE>i5qkq0Ill:hv`W&tM1!,m+eL%T]-2}7Sf_=i#REPˈ2%L0Tjqɱeْ|r*G-ɸv/ObV/Ϸ3U;?7`/.L.|mc 0 0GM,W MX5`yΨj|x&9#h&Mlr]60d$IT;nFRRK(TZLr1U6`v|+lp7WxںJ!A[ @]+ŵuVZAhT+!KN0(JZ)孿]lU> ۱wLݽ 0Fjy?h"?Ɵ+*"" h nQ4V*!0sI=PfY#D#Ȋ!+S x•e"BzAOPXR)D DN?n%g# ix4Jw?Oŭ/pG7KN+84~'V<"e9D1#c"n9FD0 +$+vO8|4y<[SK9; $J-sHbc6sJ(eYu n+ pYRϬ9~GB# GQ9 yBD%Sx)kug}k~Xϓa.C9)h8$0tPEm[v{pc^jC \tP2p5ΐ 1e:V)%+EZ(#IV#0*&:'yO˜ {Tp;gERVѵic=GG{L8\8[̥¼t u,N𳔊ه7'N go?I7@ 8ŮdXR+ 0hN Mt/jW}6vtreCh.Qx~|͜޼MG7g`<:Y_ױyŻ-h&|=P^vzazˇpEwӽ ˯0:Y'@3d;NUeSD`᧩HymuL,%t>X<䟹ЩeV$FG#'!Dn:ݗiW #KU5UCQ˹>{XQ)E7S!cq:I$xf\ɛy&,.w*eªJBޥ>+Rg!qqѫ5X`4*RwOClmۑgMQl1khF 1lOBVFw2F70|^.ڴ^ AC^<2]KN$P}ErNi;wwtP}{yOapl%dQUDrtM]uU;I6<73g+:=LxIcA@ f(K$ Lj+4ÑH r׻Q0ڳ(ED-BhWxvCo\FO$ 0w֑{ VDJ!ґAz0) `^+J/@Q_M[K|+OVՠbX~1w-½ULob( L1~&xAwFqn)9[˨,R+sn1$:]b%co%KIlXKv)K#"8/SBE$ɍ..RNq,z̝vX-"3øaFhNPƖ!nklpP/ i|5B,;1/m5vGwX~&Nhк `&DF &31GyI\N$~cR|WxޣB0q@ Jb}(*IоJyvI*Rjq101Svd/| x&li{\cȺZ+:RPӛK7˓z&>%r%1ZWޚIP…&s阷=h)|[L4>pΠA.$s 9#K]8H/UeBkFՋ-:t{E F&@;L P3j&@j(fLCV 5fL P3j&@YdL P3j&@5fԯYgL P3j&@5.XXdV䢙\4frL.>\4frL.E3h&䢙\4frL.zxdMBqBdK5Ju&i ejxT;$9Q} =O_b`:굽~˫'\TRI Dtw6 *\JCYtG!~6 \:qGaԄy]~,ZT]LnO5Z~\z|QfF'w&̕nbҳT{v9R7nb5$(n5]5UCѬšr Sև^OЃ*uo=ٹh,,twF6Vb4J^0!U2 C4.7}P(TmRX7\?@{wᇳN0Qgo}svkXq=0FԣEQ(nUPjjoZ5Ul5iuoRl Nʘق[M J??]~o;#"rowՎGh@`E.䤏Ψpu+u]J(1m@AW Ѝɍ,cHxL2+ZY pTP=Xۇ@'\Emv6LU^Y͞0U{oEI:ТQ;~zϝt:L̢FิKKV*tZ٫Vӏ=ؙm(7F4:E쬁U9lyv0ޜE^ b^wW^e`G1EtiXȃ94+šox7)4G~~=G~JBFN uCɕ,'JDZHI* j,籯R/pHu=0s3w%E^k}$P@&Zt1'>'lirJHS Pn.y,β} us7z9}fg[z4%_Pq&Sgy^l٧~@Dl/uB$\ب<0T3a١Q@:VC[,z>[Kݭ2ҘC\ͷ6Ff#%"-3|# ^ 28[pկ45|QDPXJP M MaoJY-E( w]ҿJhX(&[硷"# Z)I7b?;_E}5<' cz}{ǧw5??>O?ɷgE>#rFnNӵk据 ,>=k}ا8ռ_}6s|CN6#_-5od/@/}v^iPlt֎d^401 ,LTxvP^L`8י|gʼlve y5ED93O8Aky*d:v B.3{6&y8bhɼ878l:; cl A{MLRFB:j '`Eg,MU} 1t_JհCTpTpԷ)3Ъ4ScM].<QֱgԾߗyD޷S U, s]Z E_Kl$CމOEDɡ4uz>yh~'~?H15B '˵L 5&J]ܛH1\1%!A^s-SֹG(n;I^_5"@I)`ns14 T}t%T}ː#H ")UB1Xcwyq#UKc6aM?v  B ~gڰ3Яdyg kL,B4\d(,jLv1>Rc/SB}4 UEcZgq%9ԐR_"lK lc.r{լrTc&T qޥp ޲M%"2ǎǮdk|F~ok.]V@[wVNI&&&}@黎in0雷imLĴ^չ4&z;K.)h2$ B}JOͼ/1KQBhe& E{2%*S`RLLdkffZc5Th I*Ar.:RX reCBDu}J}OBI"_?ߍ2\*7E_奘3 VKP&R|uY3!;S_5KL)Ԧ+Ξ;suWޏGƖSk'm֟C!2 (3MeVsQ*C,}t2V Yzh\G.Id|Dc%%|o2~6l:;щ}C՗F2^ &@SMkJ=YJFwbrhZMŐf?qp*U+BjQ~"CZ@þOŧ$PkԪg3Oy_F4.H+h~.`I_*r-M&xǐqTlhnNE9TWS(M i)!mE*9&_Ԩ"c,`q[W~Gamɕ,do-:[k]jY(fhI1A0ͦ~{ t;0v\XGYȾA~$_|XL!mE)_~\@|6Ɓ޿P6!s*ǵ9ɭS[\ԯe@hU: c:zyxwcOo} IG޶/W9Op{#Ŗݾ]l~s}g.Y]OЏw[Hi]ӗUX9>bPݭmoWk/>/uqrz*NaO@ `m-ejN{E2B-9/ .S(ȐmE"8"& J>mdq2KP}A@p-ѷ/3z#H[z5HuEu7Ŕgx:@)~P)L.%?A6e*)@3z ^>o<!"`  גE!ZHl"DJB2Tùf"l[#ea=Q4XXJ)Jb77Zmt?Bǔ$[zB-r喴BӐ"y)fz .K$"8C7>)zroW AS2!M OŻ<k\g¡AP8Ձ`AeC\$8LcLpCA|Lo^`ͻfEљpeAæ<ҘxH}O:|߰Aw}q+G[;2ۡ}'eXbDIxTv ξ97ܰ6QfCw>^ד,7L jI'oކߐ)k,߬n! R3L)Lԃln"Ґy}@v}PT;OH덐%fks=U/Ys{~>y>[Kl47%b, ¹Hk 6{e,,8ؘ)%R%o\PJkX99ps{@I畛H\hQp&6s348S7l:;^\˫|z+՗Iz˛tnU6O=Ej̽nUEr) ^ P}1VP Ct9lcx7?f3b8=3ٲ=Cȥ<^80rFDE|=`vqwaq}e.Y}UﺶF䒯v;Ri]GZd&q{Άz9kUJ]kڒ\~6+ۏ~Vh D)ul*Ħ~{J ⲘRHf*4S87*$1ĦYwe0Lg GS2H7T8{P2:s DL4,RW[I'4} ÛMכrmHbzK޴:sKN: ΡiI=k[J0OcF 4PcwZbS4 ѫ݄j[2@%c3ұVB~p- ۄ#%8e8_g积75Qt@G @= x(@08&ahށ@ۢgRι*)KK:FzrEРX]GHxy =W |Kʐdj16fFc1I P2FN^x< G:#d$qo&t3ŒOE4:=)^%;Ԥ/PLQgjKpF)|8;2d͝Lsu+oF={嬭Q| &MB~!5fׁ 5{U֖G:KC6
    <&vx.1j\=RwJM1OG.%{ĕJғOR6ϰhj&ZPoH%}P[cf=p,zh@ɠV=uJE[% Kr TI+&c5j^r );NQ1hXNЌĝA1OF||Ua𤻀i7W]Ѵ,Z2%ߥԢR)HQLPmaѶ""?0v}1LD|P RZ=x}ңO*GrG4F2d=DŽq nJIȶkA?.FZQ;eXr[.ןcMB’lX .$P tA/[0*|r;ETUF7AA *0yn܍I+~w|v9V_ x;4} kN1@(N)7nXJNV>;}l nP_pj!#$4-.V)WQְ7w楸TL;w ͩӖč&ghΠϝW>HDL ٻI# ( }7:@}nBhj Z53F1X 㤒SpIKM<2i ~tkﵨ{p:շ?\OMCjcصI$:ǚQKhO=t-'wXDy[jV>ROHBWڔs'%G|()綊vmМl1ľϏu~9:n++h⦅^_=~:qpeC2U?99a[ZKznO5G{j}lu HsMKǷ{_wp N'[}쉦sna@q(Q@̊ZfEtVT8h$Z #iMrBJ~"bI#u;{L1͋s;;?z/{uݱ1Uǫsp{ϺX{ U7[CdLqp1+&LjI從j[-4jqK5OYM#1azzn?C7)8 ot5eoq v-xgʸ[8+RldeS"] k&)1 =ܱo$5GF_ 7[>;Ŕlb}=XE̘j5[Vwѯk׻+m뾺`+>r}]<g_թje}{\> O!Ba{N>Zz66zpvRNۀE#3&Ґ|*A>8/< = xouvZ*c-7GKd\:I j6 D`d遲!ܗ]7~=٧Nhk/9ZJa rpړϦUH2Ҭ "^fw5'ƞR5IQ-9+z-U5TkSrϏeEП[~ Ύ !iڂ g'}jeUd衅Ӂ1PBI׷Qv%KTϾ{Uh\Rv=YvvOΖ:ϓWVg.@2雗}ߢXu*s S+8a ,Sr%G-/iUo-u]U:-Nī/;۰tmV \Ǐ8rL"r[i.r9j !UGX4MkVZm'[W.3ǧ'Ze 9a9ONaVČׇj~ KU+|ͻ|.K;.~Knn(q#qxPW T6b UE5PvSTTvZ4{ R*$yQ*WOHI Bb?(SpMsUf"4ӥK.Ҝ:v89ǻ\<]ޜx8eVUjj؀Ӗc333r3M~'߀޵#?slŧI2`lg6c&`[;nK۱+,d,XU,VF1 rF`&1X8>r >Ikw#wcqEsǚ-~1{t\8YN.>y7Yk9';s y\2 1l;*1Z2Fx WߍO X}r|||tf;>`tz뗛(7mv^h㿢>ŋ/?xvi09NPF(?wo~=!!8%=D/~ڋ /43/k٥7v8.Wsisp`:\4#?Bpnorz #?2#婓ѹ௯͋Qu /_hr$t7|OEeʚJּ~bEȟeCQe zS0{ `8_(2qTIasµRCJ9y+mr!z6?i\HڔNY& S 6UP_t.?q7Y#g!NLgHsױ׼&}sY~c]_"@f߹p[Rݶ:j~/ݓ'QrV*ݒ1Q jQ:+)َSuh>n5ҺhwGu[VJlQZ)i,5>:-`̉$B2.)MJq1gD;MMq .x뜡G D(jN*9kƕi_^K 3A&s[/Jy 4%OڣzFR݀̑e1!$"io!Z橏$Q hͩt"8-OY=Qi,J:cmِhOv}C?&(=zg:I"WR%%k7l}{W"ഊ MEd+ {@GEq q)rS 0b]akܱ-x@؊ՙr\i ܉m>RQNk8Q+O'8U0Ë?;UB:=&ƣLwִ c%h+}xw."oKݭK"@ْK`c6w؍"AY/A$oK:W0$i$-"nGR N-/Eʔt}qavv{\E^cu_aX.h̋\˭&*jqiGKLe;FV삗uErAFƭK1VxS<52xc(% y4S#r˓6\x gD8rH<P?\a+|, 4c4zod]+ґAՙr5²LJ-SBRDMR!j"pva{ۿҴO271E X `.餈1o*T S*e+~"v"mbXTJT$MTy%"Qȇ^P"]Az͎K.F!IOAj"b}WkNWggn?<Ғ]Hr8[FcΌyads,jKΥ H%Gᡲ ԡNrwg>c\ ~@5$3Q `HN8x\`$r䍣P4\ 9OJe$PO2ʏ=!bYڧ?bh=+qpr6<\ 3u /}^6Gq@@ Oq=JF' 1*qТzosFIx+4܂oٛT۪U QuVN&7{yPBKe n9>o_gb7Zk&j,pe"^Aqruߤ(b.=JcހXdhʍBx!R6/Lc6\%n#ѻM&3FuGB~P&bBY;Rx Q0aPH0N9{& @QfPx~Dą y8^KiB9ڥSZtZ٪hFLo6;;[3Pn߹sXw۹ ݝŧ-/tvn|GIj1VDt\Զ$\}ĥ]rhhns떵&@Jǃ& CL(s\!8rhS8d)ADf CQC= wA 4 ;c$:@'-*RӮ?mZN H۔|[9|~;X_M]SW>ףY{`WvRlq/ݓtH1'3Ou)HqG?<1g-]͒ ʿOU>'╗v|Z C*Z e{ڗ4m'⽫fhEyJYU<3oKA 5-uT.o?) EUQ}]yzdVlfwp 3 |WEgpZy ag/g |0It.s-l< l  \er \ej9vT W|*Blg*;pprW$B+$X3p%" \!TltT2"\)Aw2ONU&W]+vT"\i+$Xp +pt •J(CpKU&]LL{aʬ9P nD0'm nF}Ս&]CЃUkzy²rK\v{bDu[p po(MˠG.$|SN'tq̃jBjq#'Ѳ71ʊFfyt-s5!o݅OC뻓5k`ֿu+R\/Lf[Xv.1toρZS?g#u0XU\K(ڍ- w{ni8qͻ/}s5s+ ^^!cYI~o~lÙhI{ NJ ڇԁ^$S oirq_{?D}"AF1|ٙV,Vd'$Q'nXPrj>'%8N0ZկS98H0NQBj=H>Kʓ"A.aG0SMMlތ5 H@ ǜjR,cKaMQ1xl!ZC0eԪ1"2M˙cQ<GʰD!F8bj0MΦ}x)%!|p J.Qx`ID$ p(Ir IDdw NfĘ(c&YJ ڹDmw-HXnLB!aC$6[A3@nb0˨s _+HPRkcNpmHIKoG U#"z՘A#_$S%c~NPJ,[ ۀd:/"dB17w&EGJF!@ n9ń _oK&Tu}cu̵pHRnax*g)Q,#pn`eV5ͨ K+c3fYNW ܒm-ؖ"?A·ZsiKlyAFc2<42 8:C/hk0}A,͔sj3JKv1b8db1l~Ӹ|k>$ZGg%;]cR6iIcBŠ)ԙ)DF;Zb\n- (hvN֢2_~M0a:M$ҷ hHaؖ3ՈЗt{GRUfMư1%O v1!m*%tŸP^ RiA6iy],C=hzyGr@{T@KDUW&[On}GmHqDcĢfH~D}gwl$95.0H NH?/e1N'ԟbӼKmQQLKi2}Ċ AXqw.m,5g>"شAĖ.W5-*rpR Ha:2d_4JpErr,-!Ѯc$w uPt1;RԸ^K̛\=VQ_0hm\*1j|!9tgB E䑽 q#k>h0t6^Țfd RTЏu$j NQTDup;<\4 Kb8a1* -'Z#[u~ELBiˢ94DTP"] PJbY^u3QED2ߦ@J؍o[!ȋ!6yr+ϟt_@t{4p52>17HeRcu!Q&^Η`A3F6%a_[])?Vaf-ɻ֒uK+(΋EZJD1H~]gnf䥸m5"#)^r樇r#c6e wNJDVcHtpV$rEf@ρ'69B+bU߶QOaƠBG+A$XÖb upHF8[wbMySlj NTpNʢ#C:QGI1-1`bXTuK`qKjckЏٌ+j >pȍ0s󙚼dpL-yXӗL%놼3rH&s@ኡښ D{eryӦkPiKɒ?Ajoܙy,BwhU|`ͨ@y E̛NMp 0Lr3 ^,XY2O+rJᰂl*'MkOVO`*q$ &IrrC 1.* ϋMl6VՅnKS`˩ՆXMf=Iv5|* r](A [l$Di&6 dH_0tz D9`}V,R5?.muy9N9!'9dћUhtp8+G'd?4F rn; voDS84ϝ@L $j $* tH'l: N uH@R': N uH@R': N uH@R': N uH@R': N uH@R': N uH@qyk}4߼ѫ~Էwx6FaU{'@GA0q Ǹ1.y2wo\JoոtzJH?uu\\}'9P>K뫿ߋ&"*|0xpeWa\oE@ؓW&^T{EJ!Ï?ܿApzWnz}{FB &Chm߫77ֿ9 9ګ~/h?:rfwmX^ʫu:iELυ0?~aBW|9__$aP;k|u$&{r˷owG ekwA%G1d-4ŜPN(p/nQ|RϺ{T{9|#ViqÝț~ʑCPf$.3~5O+4ĩ7WuW6L@ GO㔗A4,-'j&*׌a+>jIL&^~D/?d~δx*It3+N=Ml,(x8rv۵y )Ͳ)/-8ActiK ٜy"\Af\\JF;De⊈ sudW69QꀸBdzJFj)ۘgԒ}1Xqu@\;Ӷvf\\7 Dm{ǕL-戸i]Wf\Z]|2qqN+L~\\fӱ~)*4pYjG&h ͂+Qq%*I/wapEa]O慜iåqRpJ^j_6_WTZ+2ՇU\=-e~ǖ.^{$aSY"]D4̂+OǕT\W>&q"\AW"7L+Qv_C:(9h"\Aa\\JԆJTl{yp%&? <ٕȍJTҽ:BJqfv )KrN5~\Z{\ʠkWGU̎xvf7ϛA,l+QtꈸJ9pσ+4UѰ;!츲zcW N3N._x{:2^u*ٗWR.Iy0]o  -4ȰNq`ꈸ3.wړW&\6E_ݝ'?=A6FO_${urK~W< ;&Wf]'F)oeYDI.ZqYWꞺYZfYZaK rg\,-xk)΄+/4iP 6Wꀸ"Fy3C.9WRb֩]WlS4 8;; DJ+Qꀸ /^''#ryJݯ]ʝ]z\Eli 3\DF'Jfw\名5W]_fp`.=vtak*depN]~]o{ɖ8caji0-r/}K:P-+^>uOhzV=XHqq0zQϾoZ՝ ڝ;eMϒ!S^&a[V^{ZPb&"8SbC.4K-j{-*)i}'8N+7̯,tR(++T.pip%rWq%*Y3Wb3eW"8ip%r͂+͆+QiWGU@ۙ+4gKrNs*fM+Yy֮DLdR*tQAU^"](uqCW.-2ZQJ=ҕ`bv*'fjU ]!Zvv( ^"]I*I]!`Į2\YCFkz ]] ])N,)v8#CWn9+mWD3J38HWN%+,2\.JV(5pEI+tB1V CzCWȮ0oXs>3]B虜v(yJ+1թ]OA2A +))2\J\iPʁ.UO(R )to6vkAViZ5nݛctđ;-|8=ZM9ՈUԫatRI^B^*DVL>C'tgS((rB -%އ2J-`ܘ V/IR*tQ6ޓr YHr֙3\YMFkH QR@BDWX(2sB` +)9'% Xb*-'}q@WCWJJKCqZoWt*] ]!ZP{Fٷz fI s^]e KV2J3K+8 ʀ5-2\SLRUF CWȮD)PWW:yRJ"CWP]t%:`x_"pE-%+]j#ȽthTDQ MgBtFi@HLע Bu1tpѥUF{*3J)@Z +5ap%)2sǶBٌ,tu9tR]3YA1 WR*=׊h;zp/*P ]eBѾUFgJ-$DWXrB(&Ԟ*wB.4Jb*ÕhuW3J3Į. ڔ " )'vᖓȐB*qJmW6 +Μ>`+oΠjAWjSR_uc6CW.R*w(Jp/**JkꖳkJ85}z5 ZZ5FrhiKFv]T)x#ۥB {V=k' =[-mzH١xvGd,缝#@DMLJ`7SYj|Up1|LKtu1;x棧#|?~g~fq<1Ū S'y5K ]#D .N3S ~i=W`+M~_$I(d6(] A3ze!$߂}@V2cdXp3IF/2!o=I&Xs-U+P(I!F 2*>NG4HJ} 3Ƀ"s^~|X"Φ m]X_Zh'Jy.{Zvx _ua:7߻9> on[ϷrF2Ҁ<V>y(q7-rjOmտ+?d+(َ+Q%H~EMb4 'VPO(eBCL(Emh$pPk6V;E!2ap  бty[#AK鏥Hp3B(@*Dkh36Jp霯 tX MycjC*FWa:ߌXhJ;d=Ew12NxfZ%ka1$U vU*;j6ڍ9L[aI{BE r+O|P~簵u7],~-C R_LR~UwuP?=:]tq]}= ퟶ No?0c= ZĎWqy Wq:~n:}U7z]ˉ(r|V~2Z6U^I]kQoȱ)Y䊖E64Q\$Q5/]awKزuds\^=[XtrwlZ&b+U iW1FzLVgeF FiĘin6'6%q$*TѦ"F;c8ʒ3K~K2K %2gEr0> bBO @8t01YޢkyV唗Z)}6&'u<1)P p`e޽{u+8C{䈮䥎(/ 'g\ܽU?xfwEyRl )y' >EC54Si:ϚN`I1`Y&dzn#RA%!Q?s&1$ADU4N^1#2Ϩڙ%r8PxWPIEX:*ehqN:FX;ig]}铙̬w*~ܗ%}T2 $XJ4Z/87D DYI^$z0.MGk>BR*98J$qtB6]ݔ{1?偹JZN8XwG%lb:iDWz>lY׹W1z,ȃS>r,Wk}53 KWXn.]JdL2>ͺB,Yaj=ZC͏|D͍|:=PÞcI:{L^}{Rǟ_'?(!__F O>tn:22GO6E\< GfR3;٨Rq޹nE֘axkv~ys|3B0a̙;$G*Nw៳v[:M*h1fJ*? 92t/ufT^M|yú֖; 1$1]Iä0m1NbJʼ˯3/P3JPo**W2BSV[R7zN$, PS2)@R8δPKN (k]bk#Ss< C &N &_&eF2 ,ZT"=&b:9rhf=Yo6)zjwO{ 8 #! *V=qdNc D'IrN0A ih׊N*H\@D  oT~- 3),CqT9* Y@ڋ FU]'Ɯ)^;nzf!dgcP],6݋N9>YjG8e]仚|q N7yẕ(=:.9^x{ж~RϷp{XP.-~۟Mh) 9+Ns9:k֔zn6 ;R F Y^!O{+ ]llcwqL q ū]\-\!?"Xq}uk~;Udq@:?Y~8-ZLSk̟Mɓ^V*.X8V*> jȳUGS;UtTT'Nl23-`[Tv}SO] odŴX*gHIOA@D@5'EOы/nwrT twVSATJ P$9) jHCE'i߲P,Nܨao6ZȜ} xI3L= L҇j&LAD `NU1DdLrHDF Dd7C".θiv}Ǟړ9fawׁ"D V_ш0l/kl}6j>[L ET> - >H4"!Sn9aX8ItHj$9q>YF86+[rԝY?ߵ`￈H,Ҩ"ƵG\xpD3Rk!tj"$0n0Nwloݚ:%pOQe]4<ybIзC8L gHAxէWe8kdep29*7v >Cϳ>ݐΤ"}J־c7h[»Ir+y PdYBH.Jm#I"_v002Nf\ 觭,iH _5I=-ʲMr"Ed]U]U]Eskjŭ6MfO3G4U2(. /T>`NsFq+Ij1ƴ?x}Ϥk/Cr&df2>/l_/ډ%CzFq"6/bXsjbAiJĨsFVN䉣cU-Tj٦&UaIQaָ Su72d辪 ֢pn!ʀc,E;%֫2K,ߑ۩O z5wé%NXXr,) Tqͣ 㬶Ƙ& LACrE؄ A2 h¬QFYJy>%@ =^Z/8[j3n7R?b2r avVIEg#eZGvdGR AJa1*00 *b)+NzUӅ@Rp>깖XXubc8:X$QEdQ ̴3nA h0R*Zbraf}Q؁T@K[/6}rw8*$vە*']?onp>Vh+*"KYM0qA d+(3F##$\^/jC1H W6NN>AS*2J%2>HP'D[$Yo`GwSLI$yw~3vgZlUͫZf<"gr*bF?\g"n9FD0 +)gxa<ܷG-i9Ϗ aj)}QY%Il,@|0rQZFqrtޟdjY5qȜS}$H8RpS|ʫ.ڽ'*K$X #=#/uYm&$ ԭ5@h@6$/5Ww7Q]Sz)q阷=h)|[L4=A3 mU;8L1BL 5ASZKj oh![A.V#0jaJaUt?g(Rf]}?W?:xOg'$'CA׸pX?u2:tu'_￿y9 o7as0a^:eG? wg V6y9:4Cxi 5:t^tz~z}5{7٣[?1<1ܸr^]Woyq_Ą d38yӴɧp Sm_e{5zn-C̬XޙAjqR*唧Tkodpi @)o)$NwȺpI=:u͇Y,8|G!@]tJ_]@̈́QtKH߽EC(UE@AͪCi9AbՍ,g\}5a+Etqz^Q9Ӡu.D 0RFkCS }2PP#wn*4̾z68ͮAJm]tQ.whrݽ6;MU w urM#xj&ju3zlochFEI- MՀEƝfM4 FDVsWbI2;g4[z1^X"3:SÔSImq(p9H45ۈxF 56`<{\,s=4:#reCaBi)5;eZKY> `o⊛;/\=4(=z9W"ZmeHQXhu쁡uFa6^/kVNM?{W^$Ttһp^`v>PŇ07/-Fh0Mr?(Ṡqr`I D4JfacZJtVK Q)>B0.]uh) [GUemiSNlZ~K½y%y Pq{eK XVv~TНuz 5 x2J{dddhe^mW3RN2?@ _/J\g}Ċ즺 φj D‚D^)3MH b)4j `Ghk W5T#4NP@v/uTMgNx$vIe)0m|vͲ,h&*z `DXOF"zfz0pYyεցGETS}c-k؊a\ y啼ϼv *#EJNcR@03,*͸ic  B dJ acY צg_O[WiaR(pCdR!q)0v QH q Ң[c*xpEpJ%!i2:jB1٨D„$i#U &{ZBj[c1b>hg J%O^mBʁ7KLso |{^ZڒqIO:~:~ Y0;MV#=8gOk @5Y\aOoOs! +U)(l&YD_?(b3СEz~HFQ2߰@H SD 3K۽:^Ϋ{o%P0 *:mqkҥSkLS>_uiqb;;M8z>l>30C_ sE34,W_GXեJJT|nF0R٢9 8hb }\JoF-Q:+Y>w^7^LՅE:0:)3`.ͷN/v;gR=HMEMc@vC9Cf$AqHnR?f, ~`,]^M'z4sox9XjG%hӗ]1Ҋ1iÔBI@_CB֏YbgT߽^T :v:s@ǿ3:}uyo>>}uz f`l  {^(z [ ƛMbhM|j·W65㮒FH1Pf;nDۋȦ.m:QuH'#g~ d~YtF\@E{,Db\/<".P 36yeYV:m(Wr80,pE8*@z7dmz‚p U#=azo zپXDbh΁@#s'Y`viJ3Ft*1طڀscT!tr0\#V^eȉFqL7 9W׵9?x4<0smC,[~&}!n˵wX|K1vdf3^^nL[(U/IP;4I=[ybjLdYf(`?+QE oISBhѸw9+7-LȰZ(o $%tf 'eb48hvhal+&Y R{,]ߣP=b SD1҆gz:_!yG/C [?m)Tlm|N?45f̱]uꜾͮђz_topmn4Ԭ֠ިx_{:|tDJ"ĥ\J"N~֒ (;%X̳wR.##O&IqpiǍ* TEUAA!%$j+#ԗ]gϝF]Gd|4j=JKAgBV\K//D?n׍Rq#b8CLO/?)naw-asn\nQO~֟>`&_ɎCɎ2RCKc7ו ^%rW梅A,ʌ!hmw Fuu9gm[-!zLHjif\ ̸VY raZ Ӂ FK-r.?LeˀoxQ0}ް''gه/:7XNLY;TV]iWqw"J#wJZ'r!irKRA i6s1 /K Ņ{183v9;msg_*:ڛ/O=vگEUKl KʵL˖g }X@%2r t*yQ>,W$* eܢ^.:>V%/%\*LuχkSӄg0bTFx`Ļ"3B<*y!ZOj (y,T.D o_FY Hx”F0fpQA2V8i/HyKc x⦽.u.%OE^xWĻ0zs銆f-8{Yzq5;!=<7jK\{bѱ+|O߁}|]#t?xW(֛  ~T>=j64p)KmכLF^a>wBWҮDi|WHWƎ>]]tItʘ]B,.=+/*s)pcZt5Qzt^ :.EWK. ]MjԃU|77G78֮Rm*>:k+ioj5huJu@WwicYN῏>sʫCjuczmc +/GN*u|?[FsotE2qJoSp)ƿ~}Tthݾ4]rce@ӯYb|tz.]pCW. hUv& zt5}Ze!p&ڰj@WT@/GW&u(k+u51. ]Mfpk+ lZGDtq_jt5Qn(]!]_]-EW+7tJ u*E i 0 ]M67(CWS޸^b_}tɹoqp7ǡ hq(㎩Jzԓm.m]Orx&f~Gw`S#ȶ74=hcuDtH)?\y٭"ry)3$?73)82rtvӚ>?G#b/oLhc/ϼu\;>ii|wtI͘_y!0ZpG_e[uҪhIu 抢RێgۡןWWATY]?Ϻ/ w7_V-CQ~>o<+lCMԠFk)ȷW®j>kO^>g7a0錃%;i9+̹[R̄ˁzxn7Vlז|}Vlɢ]賹 S.A, .+)Rnc]JyE93k]Ө}8a{/''stឯ)eGiSGO᤿s &hnϴ7&x]g4Wz[)h͗?|:s>5xn<..7ۆ}I!߭o-]'g/n(/BfW~e%hM #^C^>{`>_qz°l Ӣ߯s3Գ8b^"iK$ J+Q3J/hMbk04>u}x|漷WpiTK4BK(bE !;sGf?']?bɗ\b_KCH2jWKi6F2̵mimAE'~ZhP[5D,& Y!*9 oE+)Dk7wGbx0K\Zd$q- (9qNZFN屁\޽p I-Px69v@&%$i'ݹrb>-(8hZCp~V脤Ju=QK.X0 D4!%_=m-S34❣p悋ch |TTa2!| 1g󷆉FMòX=]cG(3 34%y"iOr߲s%_q2u*L%kH3'fU,ys+YU\/Co>LsL F dFg@oԑ)%&ҁх)w{ҚcF{2&XKtI. VCI1Q{X R!JXֽ&#J3\`Ok(YTg3o YN~>yT!ssf0fbQBn[:H #KG ў;kե ӅˆzY*UHFL$X &{4:9ӪYpZaG.p߮8  %*YJo XPBCE 1䅩plSeX$Z.Ka*XlTF)n>D5h_5֨w˽AAIpWBȾm` mI֨uhc$`ܕI2)me") HpPRZD2 .P_\l F҉b,qdWb%@(R +Z ȸCæV#(Ɛ@ @i!#hRA_[Kƒ 9  & b j`tZ,ppR 3(-XK@t*>*@A.[*ABiVz bMpu%t4GQF `Ȋ3wB3tO~Fwv 5ِ|1fPT)ND0BKwS;v;|竛s.񂷪`phY7Q׾ L0,͈ Lc{T^DTd$]͡Vt 2:ĀbPl;X5/Y,,tA\  ) >@(&5 *2cKYvbey`tŲaϐ2s:BTmh? ,\-d!:T?wFby7¶MZWA8NCo}tɻ:YjT#י`d\b=ѣeDG$%{ 6ȋE}L! |R GaL)ڽ,{ZJhr5;7cCXDѫ)VkjwX-&]T3jFPXZQDy1`\l%lzg1#t% \us2$C@&IPztX8a1.;ǂpN"n8TZ2ZՈ@4WHY g&;kj3y o k*dGou^Ipc+XH$I'v> W44st6vy|[?zq꿂Ojrn$ !n:s6chdmSI ]-2Y]kǰQ3y&  2PޞdEeF٧Fq!)a!/Q{ T9P.ϰ [+&xѤD+ҝ+*Tٻ$`MM;Ƿ0w1{:h-z=*Rzp$fyɈZ_wW]լ[(y T")HOpGҀVCwYall2 ?Y/++&U^ B W;9ilVC2p!v*I ecx)UZ( q#>S500ݳ5u,4.cG],WЏNc).)3)Sw{jRCuY_p)xg&LD2f4.5'k&ݲ쏼ʼ~`#\k 9N%*d! ,%7L)03Ƈ( ƞZMzJ884qnA[|* w9mjIJKU/:a-,`vLM*) "ICdr7uA%=)@Y , \c3uQ~k>]`!ީ zl/ 0<2a7^b^V\ܴVM*`}v1‰z*h(Sή 8}gH<<Ǯ:. B׏|>}fӸFoPO ED/K$EULMvfHϫ]TiN4/j,|l˘D/ J:_.3.xpϰ?cZa^=l%fvqbZ^f~ҾTnJvm#-*Ve2BWQx[ $~&)㞘z2$GS8uọn@|zCDc{^;K!ׯ) n|')H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@߬H)pα…B] (v::)G t&)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ RhJ I q G QZ{R\@CT #%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ RhJ J d:%&%5J @i=)rF)IJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%p@7R-SSMuz}N$|ElZW@ ,$8}և#\r3paRKR . At^S+ Ӣt%bN\"])ow6&B·BWEUjteJDWXpγE2`wBЕڇDWX=b۱ z ]Z-zOWu;p1,==]!JE{ <̅]yL`+WSU']u{nh)#ʞyWݟ{0Oq!+Py(t^c}+=DW+ ]v]JmE3aƔcNg,:tP-'LFb/zٓhQ[sht3p'm jt% b.B†BWVYz,ЕƳA#: *NWZҕUs]᎛CW~B+@Y+Di).+E8 $p Z4cv5DqR"b)}0tpU0鱈k'KNWRٷCW] V9U7kWrv`md+ށ8Ckyǣ+, ;Uכ`] ozyD`s˥wfwS^MguEYm2i{n>/| lP^@|9 PhGvte5+6VbSLiߙkG7xC / $Z^\f[[l9baQ_tBqqU8թK"(1_W60bf8OHSNt ])UT2>^d٦n=p =O3 CnpO '2uC)zɡI3nd@tqKp  ]!SequC5Jq Ptڨ8/Ƒ5ツ}Z.ry#.^%6:~3Bhk1DO&VIsQ(pt6E5(xb Eb ڷЫb(9zlvF 'zuwXgwu tNY7UuAk>o$﹦T[Aש%C}Z-/3:uo%RP'; /}@ /؋`^ BYxA}Q*Z'‹qSd+l .t(5Jq]H.8 ]\y(th;])\DW+-Q Ń+SxD= {pN7]\gD(th]+@# ]YUxH)X` z ]Zu Q Ct5@r1wM8kWh}$3Et5@9w=f *wBڿRz5f  >e#]uB+ڻ=Xu+EtЮH+y0t1-_BWjteX{+)P pkmvN-NDXj M#\cCi@U"JaHpI ; ]!p`@+OWR(ҕڐ+T8}M+Dk} Q:Ot5@sҚ,5;+U(thMekٖj8te`DWK ]!Sk%t(IWNsY>OBCt>]yt,hf xWVOGg .]T^5R&7q­g:NsD21<:^Bܷ4Hlm.D߽~޽T޺޹[GM^fNY!Tfi2W›T8cf\^2y/`|5T{a<[U].w->ܯ^5lb?L~HV C`XkeweC}.0j˴i-?7|~@l=)VU9`s}nn.i-׽M}L({R_YO6pxߧ')y_|Vqsm vU}ynWEkL ps0M">h q#? \||v}7Շyџka0MJsP\*WsժbƶKlϏ@̛X8'c4NTĬߩX<9P8PHk+`U i?-qz]c&MS&)rVj{_*nJU/L;R5Rל5o: Q&kPO,Qsswvct<ȫJ_x˼XEx{qN>?|E2=v=g6"!{g 6]KsAE짭ZTQr%K$2% ԎrH>I>}! ,}G_@PLRp5 -/gP0F=} obҨ$"P0oYAA#pEq6y h iOʇ tL$А3`5ggJZfqT;Dg1xΒAPc&~v'/Wx8BS[ b ͍tru_77)_) V,Ҩ#\")PW:X I&$"Dq65z{;TݏW&U5{c5K(vɻ߶TQV˗b=ܥVKnʁ4鮘7\z~a׮@Q^o'i=RgO'->;KZ문gBd)j3!La 5׮Пcqo&o -!.`GgNE`Ū5!B՟1*X`U`4+]3 jIEF N$$xR[B&ră6\0(3Y)<k1 /,4& 9 j>_~pԽ--k T -wCa'ie}y4 bb<٤Ht9~h<agoAyB ɁFEC2RI.R3HR-*U6sY@G8G'_3GT֎qUp^L[-il&n=Tn೺}1q'ST~ u=# qfqAb z wM-2_l!A(\b[xl@5u 6.(!`jQA#1bVI먀*-ĚFkBK 90eþ{8K_b>"®hMAeXXtg4ÞE Z_ P@oBRo E*0LR @w@=lJ͡Э؎V U$)".hR:@j '6FJhk]b._< FnsȨykJp!! 8 #!IPW=qVy y 9GCf)MPF Z3&754G]IkT y1h>ޜKt&TuivZ㿵tTDn宏])z nW;6^;D"Vq"qITG2ս3fh=&U:]kY:7HnӶ@{ytFy.rh)\ciM8-efYl8p eNFD0 IoH9&=Eǁi&b{vx%yGHM=R*@RG0kyո/kT=:^w=PIC*,: Gʭ2DtLԖs}8KR[ɲu6vcj jj)r x LS;㬵116ŨNZ@fWɰC%051RHB< g\F"F ti Ѱ1q`sXȒt B+ KO;[׷X&%˕+buw^OBu4;3a$[x. 9xxE "F[ QBAQA4q @1q .ʫo吔%rVRe#DnH"D8ͩVY"h$GP uheʂ<4/:^fJ{nwǯzDz;B 9>8A:.*ŔPi0j S*9¿F_H~pp#0 ߶j Ȫ}n_/fnk^lm=`N:np*`{u=c}WH qN a+( F( x/{F|Vo{*)cT ybEHhf%ґi C!ý61դtHyf;9ӏ^mp/IPMO:g?_S"cQ(DGFR:!D:bA)*|6\XdkvbX_00hdh"WJnNp%8!\렉 Q "{f5Yw,^& -%"yvսU[lΔZ#-<=B$W 'u{Z#IiKQWYrQ4cxYhZZDž(w\"hg HIp 鐦ʇ@xPvH=0T@sJ <6Zߘ`)}b42M{]. 6B60YF_w'r\}CER{̅i!9yIKG ~ 2LTff|)A|LyA c4 䌊( 1X0˥3,w  Wce.s.ezGow8'>-?Bu,N O~1hy>jS}m.er!~}`OcX\>/vPA:Nϻ0;~1]u3/]𿗝 ;5t1_\d>0#t>w}۬o%%!\Y~by4!D3dsլ2~sI.^t_dph&@=RP5&놁?Xg?o :wy1$ o0HB$]oj\~]U!%;UCsbT]R7b!#(UI ԋ9PJa<OtqԱ㎝~HĩOW7\oZc|u!D?Ի] _9zaS!U/ԕOo^lEL+ZM1AKFox1%BijP 4vqù[{f>Ig`c5,Pp"~p,g [>co ۄ)oc/ʪL{jÍfnU;ͺhY&"g[7ꞽ7͜Q,šP@V(q*lA ZI$]'|y`FE5@7`xi}F?bewe ^T?_izl~ՃT\wT٭;Ň7?>LZ0K|}\ yA KG]:8wę'R־8&xY*dsD+袭,![um'%_3#NPmstMUWFgI}}orn0PauOFrj|Ot'l{ss5GKˣN zqWi-R Php$x]Fe4D8D~k"CPMDz5Q#UkTAg XTbZF-mDXvfI@rh}8<-;QBcǗx~Uy֙Q(̓qI緷:g[NɹZO9ҊdҙG0p">vC"pE>e^nG dLϋrŽ#[13E}nVJ8j=!Y@Jq0\ W޺S~c(Y|J;!7g:>W3 jW_eNN}kθ6\,] QD'pO0*?cg֝iZ?+կ3?W'˜Lx\~uOϦo7_&㪦k|޵B? U0y} p8"n 22:|>D=Ȗ4ƀm3쮞uUu=V$M:fRwxLXfUޑ4DO<ŲO''_zv5|hcUͳ.5kZ98iVi)!m$NBbF<*q_t|àx`(O?;7ˏ?|\o͇w?|G; @«N$^=`jSkjo94V\o3yog\qzk̎jmH]$~?;-Clrrg~R9y_!I?wƳQX q+J0~ϱ#ޗn=xɤyJ9d_\%d#-/ˡV9o!aJ2M`+MZ=SJ( ń-YTΆ%5O{Y_syJj 6Y$>F :eP{A֥&z g:UtzS'>';_G«E;Ԫ!lYvJ+ ,_r?AƷ y#i6>x\ wD*m;n-e]&+gՏQTC N%-"e̘ u@m2ұ%`uk'N4ؓĸYǃ6*fB6*Z\jla}ToOI.7>quo-S&gsݩ b6qЎI;D,4툶$-V65TƠ'&8~R5Rr"e&qR 0xK!uZ6%;)=_o'cҶ{ D\CEsIFkRX\Z)?DO1L, F;U\/7.]^UD4}$(ӨpDˏi2:ʞyI{BLht[ttii2tx| os*ln|~MiZs'T*ֽ'=AТNh*d͵uЬhR1s 9`v[ n?idz sQ0}J)&+l;fl`XC,hSs?b$)_;vD5P{:j V !Jf2`9 9PN$'NLBQ =HW3*/B&͐d ic"{Q$aK"d|Y :i*a5r֨_6)D""|@7q$ԜIdU'%83i,  8td61l Js5\pGH2Ȓ@YWEks?"^~L=l^Oj\+.ʸ(\pq3d(!EM68b>EfٹlE.cy>%;|B&hgt7m N б yA!C&f $p둌yA{Y{M,Yz]z3h8yz[cB[tTu"kkodlh[ [+ю }};i+2-*0VF&ՆԞR$0M "[-$:/Ax Q9,]2Id}" Q-XR<@CPjlM#Y͎q^+J+C;cn zeS|ĜWb}2Gj`$ &M5 .Z{E<pj,~}L|/J6weQVŐTQʙcI-PqbdJDaIZ[&U- m٤iI2gee9FΖrֻCkwQroJLAY`L3O.,a[&s"C,U3f92__8$pːqs &'$g /G)k 9$&dUO;iION%p!eMRI#eRd ;t)[YJ62ϙq*ѫx+׭k[%a;Dhri{qL2ȳO1tN鉩tD]:+Gh+b׆^to*Uc awIy'e}>KF[;]Zg8jNO*ZP,sS?GEX)cZCv[Y:Kt^Iw*1Ҵ4ˌ⯠nHe@lL&BJ؍Bf!c-g:6kрe8F:-$CG.{apkM#E<me)T$BdeHEȬ˺=r":eyR=^5.ܺ&{y)bhdBycK͟/,,w?1آV񜣋3?0 =kȀOMq5Vi8Ô$%]<{Qq:hy i~:KG]pv}\qVs cO0$JV9Y+$0A;bk&G`gӏo8:o4i?J`%GB^)B@{mCB1Ӆ\ZeaP)/ /'-.fn}h_Xb_у-#G^tygӓt}H|98~Rd<O>jšǴX[W!*>9B%㴪:* =>OO'yݿRȄe`NP6Cn 0xK' u2{e[-o)7+m3kgUc] )ؖ G+7hE.7, o*܇7yA+4Yc (`SbHc(O,QeМ}Z?MB^ۊymȃ1&рd  p; Z2`+؅\bbj-B MlI 1)spC*}+` > +dR: "}᪐+ WJ5K+P1{@pU^;\r ;*ԮJW/,k z2=.]y<ѩ;-ߔt܏^z9M ۤebW4,M[Q$ O2 NOG_}߿^Ò[!o!^opL\:+OwHWw玮U,;qV!HsNV)pFHF,5 L&3s8<r:!v;i]RMy)`]\# 9q4@( NqdzI6*fB6*Z\$73 ++9zެweήoi M#L/Nb_4*Yh4iq%1ep^F׿MEAlW?7?M>MBHvˏG/8תvv ]X޹Si?ONf\k-f:2dhtvHyʹT$x KP^B% ^%Ђ,CY 1/W:n|H;>I0ΨlBkg?{WƑ q_nsH]Z-`?$d/w7܇;WE2$%[ _EDR25(y 49=5=UOUWU#GQ9+g`9"kö2ؑGt6tf ͢) BQu`qѺY\r/RSTY,.E4fVs'yr(`d:0t.3pݬydݬwZ=q:ugDa+jfhdGO,8h);cŬR"y)BelV0B`C:h :Ymt >8NF GBkduc<,kq+R'o*弸b嫭JzbZ6f]:ž&5q2ǧ ){4I&GXdҒIV>K 1(DZx#Hbڔ.0h%H#8eu4dq ɘF)l{ \֥ȲRdcmZJ8 "bȜ[ˍ6Ld0l'ijEC-Uq~F~~[O\b>$[ZRWѠW {, #m4Zynimg]5cJِcT<:e"!# ƅVQK"BK.!BLՎ{ ͷ~yko_Ȍp[$@rSHV E5q$qA\V%1LvKʘ$3dPַۈkI8=McaCheUbWkmk<}YCǜGZ4~s.,{⦫Թ]yy能Ţ7e|A3^P|f){ 9Ze[kHgBBXZ %PR)a.W;z\]CPV:Rydzl0   d\,Y'  `tMy Ѩ(0Iɂ"8ƀcB9#bPDQ7, Rj|$h<3, 2,4HvLBbA$ԨL6A²a=zF7v9w'Bz1xR|zQy)>i͵'$0HscLF0L-k}E&xU'7 w|nX{Y#f번kd\T ۵ͱY u -vU~GzoMv'T+鼯j'+'H Su{֏rώS|U%AY>dTJBTo'̞11y(ʳl `-,p-蹏tLYv7s.?Mpn_~^zyA| åym1.8 (y!HRiE朑iAAtNV#E#^*V1N1B"bDaIO-y!4aL4Ť%3F3-ns )H;" @lHLGzLH>?,jA-w3KM@]fV<:X370 SuLyq`B2Z}T,hO3g0|lT=y}/; F%Rp 9dyI&5l g[I TmU=q̚K}悛C2^byLdGwh4>pb+Iº -6cG,PԴeH  &xY]Ѕj<ix"m-"e1 !f/HI AhHGF-&v;]$ )SND5D:EDC".W2Y;\0 3;)u=.ߝ/f?>y' ݈ns!bqR\T廟ώwu_ۢO΢㜞@ c+@Iz.kUp֟F z'ή~ignɈy B:>^tqǼwݝgZ虙;NgKs^s-,prKx70䦄jB;|o~7%]%ƒnD/U/VءDhfᆥzU?$jӜn pI_u{gg-H7fF׳ûaܷDEN8Eʥ7Mc H*F2,TTvاlr! d}ʶLtf%P)v2]Ө&7}3 NդWޠrZкtˠB/!ݥQ̲֖*|0V!{aV`K}-\L qe}'i֦"MHGpl]B̌]>MީY@ms3Q[oV۟glY{QΨM|z,)=we`6`gN.%'A! Ht|` h2)IH-,F$LXx5S,*n8 u,ڈhϑ')CsȍH$lV,BoCY`K&ђ *F^?T4M^qSS^ȻW;U1h:|W!T[˺m)Z*eYֵ';nNw\ $X=UT~7S8"tHݩ__fz'<FpaХ@IxL#p%}c C[<(~E9-iF'Ҽ<Ѭugڼkyz \Mg?| c }s|:sq5ϴo1dlHZƑZ?aa82%?`ŲDo^v'[r}Cnuó(o9i\i.!uQ(]_Sg;Ei tS7~t'~vO/wBwxN+m#I0]vy0fnt1ky\ˤ@J>f1}#HJEGJ*Ze+YUE/޽z_{;.̻ݛϩ7z)Px foFawlqiSt奥 .Uu׋uu[^soÇ3f=Mfa$/g_O+\x>Ie zQ\~̦0ؕ/GaqvV 1b2qap=;h'^2vҦ@/MsKl#U`^6zyb`[Hg~գ !q혒TG)&h)rlXB^F?4`/x*E rD.>l<^Pv饉kÜN9^-}+eǶΏQˊ+hհl~h"KYvOA~݆4WX#i6>x\ ϯXq1 ZvZr˅Zų_aͣ@+ʶu{1Apm2ұ0ʼnS"Md࣡#1x>i2୍(c=R72Of+> VI(J;.rq4j^ܳc{-7u;'vd`uqZ66㳣1@ 9$NT Go,zPu/ءv&Oikk:Qӳ-vi%$!caOIz!&9Bрe\hZ킒R_5 Z{}tv[iTpЅFߧڕ>3 14Gߦb}ہu{7L>L'kF|Z$li\w[/NR>&;k}}\d֘Y7בuX\Y*7Pc6w?Tr@B/]yn=-Sbyh`zC£giɳG'YG1~I{luR||>o?Zv]FH:-b&ɲr%fgm?uɜ}-ycv~Zp#eEA u_:Uhz omtz,?-I;)oavtkQgm:NtqrXQӳdށSů?:eȳn#(써 VOG_g'qYP }_=%U1~qsw*vg*dZDvlTк]5A92ab1лأ mIάE!=u,F3'x`TF5s6\t-dIGv_BQw8.5dgm@!ʾzr$EG hP >ڦsec a@fنHy)&' <1f&ԞséfTS[(&LOeV5?9tkz͂Zw}69SX=S+:_oPyb9U6 A@0^YBpx BpOUdTFL)cbp R(s:fEt)EI9R97#c=R yƮXh+c -J䜞fF;Zܼceua~]O?^ZyJ,CeA䃒I0d}Ѥ c(rD]Ridz %sQ0}J)&rD]RL36 !e̹t/;E*0A#D)L ,G|I$I))1*aF] ky$fHSIec|Q$a%s>f,PRs97a/` "V]WFD>  yH j$2 aI :L̠XAsK f_WpƔʍ6+e r@Ʉ@ ͢_ְe͈}W }%⢬bx2t3JHQGMeAА<3Ӕ 0B2Ye̹0mb_wnxx vCxGni-hlonX}Opoȹ8d85W*?"o 6<t,%Sʏ؞}[$t d˃(?2rͅP.H/ "hɐH65"Jq^uЪ.,耳«@TN[[S2!8B(xe"\͜ &p'q:7襔b>EfٹlE.cy?ʝj(>! HMx8t,jB^ Hާc[hؓ9}'hF;P XCfjxPc63L@zG,.~~d^V\}lO"3~:OoG]8X|j>'2 !~V*| G-OޝvubʎZ"BԦ`%ƅue[J2[1^ZQ$' gӤ AQm!c{e O#*wvz鲧]aǮnov]fwu3JNV@苀Kw+%T_Aֆrn䵦ͧmۜ\rm2!m}wjPKk}qN7t)kf{BjP˥R槽~w7|Qo3_hy?̦kηAgCt[:*Wmf7jZc(z\ pZ9AǛ!7mYM|òc ȍV!3?ΐU?{HmʇȻ ~@{W|Qh٢4;#1$I`2*2>,B|JԳh$X5^DE$%<$ܢә*yR^z`4!>&?Rz0uz1Y[y;9Vn?<0؝]AK$O6&V䳎6|-%?r~VК&uyu?'XC,GVJr?0@nuwl_mIN;0mEKQkvu> ls󰛼n}[p\M{7+rkzըWpB~nT΢MWGNtIsV Z60& FxQEuu:+`F,1!@z"M :sR&#g*\I0}{>G[ BGyq2]{ؾsp/JMgޢFX?qRk~TΓ:&@\U,̀"sHiZ'ku@ B9 HBِ6 "R$peJ!47`L2'Ԓ7H:$ fx*,DaU^nWS>P B / =sܝ i|7kkJOMQCer>}O <:r*(Jo4w>oyo Lg ;ڻlt0m pt! AIзz[MCQ(e7 Ad%} oʡƑH;"eLqRcA@i#KAxĨ$:)j0xaJ7.X%E!Ә 1HY{{aD쯆ӯ{Tu3-~>P}_ٿ9zKG<`Zo^=,i;ߵ|])XB(/q}g9^Ωџ/qO[}~Ʌ^+}(6GHG7o~5o;-ݸg׽LJaΝz/.&߼(Ag0jJ:B`rgE^Q0#)zI -\ 9i3˺Vh]\OJvV=?R7ud҂9-$^ Y`Dz̈|֍R?#͑O2xbLW _od"me͇bQG0突Ao)X< ׽鸍B+eO{~62KNդPQ f\/v6wݥV̪ru%vja٧eD7&r1k4P1+ӿ"ivL"zEG4"ٺEs3vکبYS2{L.E~[}^p$Ѡ Z0ik*2i3X<ئC@Ceym杺K4+"UmW}=JϬam)rLF\J2l/m,q\`jS^f◻왷yUr6Ϟ.6RƲΙ0$3Ef%hR"9w<Ыzc~opz5ٛsŵxcsŖyxyj(½Ǫ=|֜c(?8u{8eNY+NtSdyopU&賷ܵ'J>U\CQbWFktT}\ Is%\18g ԩb>?WP0L- eHw8M\"ќUzOS'Gޝ껆~66ގFXmN{nl=zN7%$ͮ_ uDm}2u4ʥθd q*ݝ}z}.UOyJI_`IJ*L{BU_`WAZ-y6Wߧsa;럦]VD eξ }^CD`(RrJ=J+".&`rrJFed9 %VLZ ÃYs-F0db CHчดEf)99}ށw*hD0 ΃O_[[u!2Lm?H}r/ד J掚/5`YZ϶voYv·J9dxD2 CLr`2D!xC"ݧ*2jkrJ9V6贵) J%;IJҤ7HϹ>HFjGzJ5,b;?Ο^͸gvݵX.nttw@wϒ[dXF #8g 9Jd1JRJ0)*{NEQ,dIP=!mMІe2d1벐9bfʈ]ƣbb`-Wӡ@mCGPK`Q%\A d IU$yaVHԐJi$[#f(PH >dT2V#g=F b5yeDD< A"%AB(Td"/A!f@fK%2]&]r% FmJwYH ΄DJ, _ ,i!I4 VFjl~ib8ټOkոd[\ʸ(xŇ8iE^YYd*,( BZ6& Йc.qǾ!l;5:f!H9rjfnNpgiY?3ݏuzBCaZޗNZ-#iFv;=%vRs ܒݺ=r ԕ\뎙َ:ar~rJ('ˢФ/Ejmh , 6GJ-,JOS |ش]^u7Q=a,c"Yt ˭M:*7rAܔR! 8,Lq$=ȴ*.t&srH.=M Ӆ ܴ9Lx9X:[V/jǻr欼ᤦBBd >ߍ/c8 h#g+5J5į^LΉB`t?(Ξ8=ž=h c Ih~(,|4Ѵ4hܠi'j7i|vgpzkrmu匿FAb(-7s&Yăp^ZdSjd+Z(F3{$!%6> 2GAwK!XÞN'{:"lN?yrCykνy<`Րs+ t])6KymoI\,PVnr/޽Wyo>N|ʿ e/D`s 6dk7JxA?_ޮ,)6?6op9?%r< S)TOLV;B(_?_q™ڳY%tFGBYeE+BV?_Gsrǝ/Žxmo09{1^YǠ}RmL=!E 3[]uJn &k[wWV|'䃢M/Yus[xOm|cvϾ]l÷9>r?6.g}~u|Mf;|x>L./Ù wϼ˜߉rM=7Ƴ[9@~ )c95kNcCX8i;RMI2wP@@bh J)+/A D(@=']j ũM J!?&Z\"ob/J*eM% !'ᔳB*'_0Iaroz!$TVΣɲ-'Ŕt7ctG:6ɳRApo" hkY94d<j}҂(B`54\ Y/M}CoƘ'B2*(| m$Њ u2QþZYwF9&\֙Bf#ΣRVe c!7VK࿱JCd=Bv(:Av93̦;T @w2L]*4̬< ?QyVTJvu m֤Œ8hngl^z>)aLgS$=[+DY:LSLS=xmV̿ 5 U@rZD^]\Ζ;ֳQzG&yLT>5Y(TEP#gYI :\ —lE.hܠ]ksC>\*#\Ȉ2hiHEu3} 5nZAx}1It s8M6lr (*#/Fk#E VJFѡz,Xӧ~͐}:zA66^5CvȲ\b@,d#%0590ϯky^򤁋1u,zdnT#@0Q9ėt(K:4%2fe,`mA )$ĸ$ƒ-V2z dLQ$F^ttӺIP 4a6F% E` ]Tx3qsP/tͷ>҃Lyf^فi4;ຌp$͋}}]2?X7I_ VackݻYjmݻ_o7_9r6gհĵp_~j f70?jqvt՞ko&m|9f?73Jw/p'g㌑n!Q?{Omm yhs;_- doe4lC6+2p U$M.o208&VH6R{-DcDN$';٤ML l."ɤDkT$)ٚȾVEi;9Ec>[g<M\WѤ`|xm|gݑwvܺ6K|nwMէ߱+ 46KPڑ1h"VgM`QEc%6CE"Ay;Q޹Jْ̟:]*b׺lC8H/ߤsq̯݌k>>~8W=AbqEd%(ϵϑuC6XeYyG5h,ۖu+Bmη[+O/gO/- @U<Зn"ﲩBzԫy\پ{v..&f% -J=V5b; #!-Yg7XK]ϽsFq*-1BBQSN7 BrN pk#׋o]k4=VIFZ!X299b:HwSNƀe@.Z:,iPHEFyP;KD/:u&Ύcs5>E;^_MoB|>sw 7g1+qvRs:\]D tŨ@0 C_*S^2c H1L4H ;cL&) >k.F<^qQV1ڳR(.VL5x#zz\EmEs'_C:m)fgy5NFi<}~Zu/13.v6|@lK&NZ{Ǿz'V M5\_2PIz$,;F0YƂEti(餰-tf_X=8MM >4[4<i=| HSIds2jiXd ظ3]2RJb2P]@_*o*ot7Bt/Y^[MSf!YdV  ^> #1SW};y˜s/7&[6^s*Uup;&$%RAiXW31cL#*R2(s%SfJ T ,Q׼[ڂ̫Ixw8첂SI{>K |W5ȃ%D v5ʑ7Z܀9P)V44Lt֙)˛hfUJJQr6f>d6AYL4y ɴ6 2ya3_lmkU^n`NVt6:OJvE5Cፖ6;x3c@,CtoڱuU݋xlՓ|/_lJ7.󹴅b%wtzrɣٙ6yyȅxiKY(Hѩ Tӄ>Xڗ&~~x%9;ן;Z-] lC8DrN('qJĥJkz߫JQk1JkqB Ld ž*껹R4W 8!sU֧N\UiR+gSrY`O]UqWzn{gJig-+R'?sʓAWUZ }7WUJ1WnOT6 lܑuGFWOV ys4)헹rO0Wn0W~zY>vjFh}_|>v|vtNj~@A3\^φ&~q`#1V,(mc8!3]&s2fRkiR4:gO\dU©*K7Wʒz0Wo\i'A2WUx]*.ᩘ+R*4WspB&N\Uq=s]ko[G+fՏ b`|A?-2!e;bV_>DI$EQWe %yYoU9UMZ]5)ாDwe ^b^jCU2 KtW/ ]51/]`R/]yZlu{+(w2YSK~Ur|vtvo6aV7H2HF SUBTJn<n8/fCo۵ԜZ?i/N?VpY+j;xuV?bu>CIt*ŨKM AȲ36)|,dinQ[)p3?_/.&I{{zKxfy?suW]3Z>,TSwۋTyUѾ5DD0G tARX\K2%9eJ!qI_nImԬy5gX+"0cm.|u2{3L[ճ~yuҢG[]-qvb~c?ZO,[/_/X^ۗgCfP^˫C?Ͻ]gdi7ut=-!1$ (b. j FGF:Lmc4&=H/mUx.fc!iF yդD!fkN<8˽vMɲbdx\tѣU.JB*t&Y]TkJo]iQsrq>ϭwidN ?y-UΖo2isW\v}Ω=ťBQ)蓧d*ت*?v$KN%2Z+ȕQVKla ؄e1I3bHXr!TA{+Pe EBbHП||V!w@_pUbAREX1y6 { Eļ<ꘫHGfN (8&Di祮Qqhxɮ"Ȉg;M=Odf꧃{oKLR&H:XMȡutUQ̖iPEZOl5T. &Ч \Ϭ@0tȔYK*YFT(:o[T  L&TMaZA|y>'=a~4]~w`t%R&+-ċ3`d(d&9Ȥ3ep8]zy:dz;.s.# m֕}E=6r>L/.v\o= \G|*kל[~lgYLרw#R/|Y#٭=̀ft9;ߪk)\TNq29u!3_5H%ƶ-ʶ"mi zHiPN7*O?+jJ HFGԞ"emr?=,v=(F;?Dճb5AQ*]@s%(L4R(hyL9T&Ka^eQEv.][vkQ*IYT̺`Soܽ>qM* +=E)=ղ9oO/vwҤȓmʄ`l5؜MvXm,V]m'*i0^w>efHmIY]rAO$n+وb;B 4ZX4Cx1-,X\%_L&:nMJ3"[XkwǔQ|=ag`L)Xt)(BȉJWHVVyLI:oJYe|<,[Reg4r#NPO$w^?{seۣ?ߌ/>$T47%NV *y5g/2Fh{X1 |V*žx}^{[- R@Zh| ur[C u GT1E :մ ѦBwW8CRт:,dٛ׍ :Vǃ@ըrjGUF!tYĞ7zg*`Rc`o0}:/} _צ.cֶ 4͇u(E}?X&I Ndlm0VHs 䌹\vBL":BbT @hs kbWC֡RTk7D6eg{n ވ>;nyuۙ{/־])9LH'Gf0"S40GAXQ@*`Ei%qV @=h $K4Zɑ q-sJևxq`@_,+1c$}o_9u[9C X6~qNq`Xpu,*ǶeW L T:يvZ,*j;خ^7Pעu?ܘէq灧#5/*Ns[\1'Kv=:4) #[;rZ!`Qe5ܛ|H^Rw,_Da3i+eD>aI.`˜!i/NiBT$1$3 3gv>zjO>ez(oݱJY媜j+ZV*\=E0C+xl0bGՎ7xo/[Ʈm{ ]Dt=ùE7ۺ!soy}1*$&-"[9-n|6GDҕ3=p|8D~!z6U*|8aƷ,z&HhjM.D ӄ!w4|_VxՉ9JUB+^T9t8 ?d{fJ@FI¬ӪY `S&0>'˜/9kfrPСhl0B}\@8 $uB9*0}Kr,ɎW;#gCaMc(5 =YXSN-ިB2ns1XcA+Y@HW@:Wt<ⅇČ̀TlNDkyt{dQ1]'cG1-P$* Hg|4o ,$, {E1ҁ˪0ͥ T^{&yf^jL1ry] 7|5=0Ph;E^qA mRABK- Q uJ`ub*X.x즏G=>c.Lg o0?]Y9,}9`+{4[ܥG{^Ul Zg  ~=[CF|*$fJc` +nBqÏd=+,C Ԏ`ݞþ2B|t yV Zm aM9SF +'qK|b'/Mݗt\>p0nr؇a6,}:ҠGsL+@ȂRi>hc2zV@|;)#P6!HT)jm8eJ!47`L2̧%3F3 Q;2HpG9^1G]||蜚FwiWPgdž.(VZX Ls磡jSA)0HAtxFt0>V|!;vk7R(|h <4OWP*8̚K} LʊYt^+6Z$mˣw%W'ΝN6D!}:H}gR{$4-VGCAFQמ]?Q6oޫW?=[)XB(౒ܸYK|.Tp}otGF۳B1zV\uom} h49?~ۏy}w^?֍Gf~}uunnr?~t ͂J9N4;ΦO0PFܾb7L"W_8_~kK4xw ѷrfye6o^'3wwq{I~yjzN4פ̈́mҥغiVjMGænw ֪.n>#7Jx30OF.s)+gxtEmơ'jjS^fhw/޳WvAWxHi$a FJ>zƃ.T*'HdfǃDr^Uexc+4Eex[auh[W薣Ԡ8`}l"-Ao->7&*i.u)TM%WUU.sEnS9|h3cVt{An{Έ(yD+N*Y0Qv "F:ͻlr&+^ĭVyx!q]6;]x IzR@J{S夡tlvF&9e^Y㹩<71ONz6,ߋhqh] 3yld`9piR?h|f}uL:^ lZ:^KU02TKNp^wRgz{g=^^Xg΃* M/ThJ)T. 7^4ע[(ϙXf.$q)3LH%xީM]9茜 1֪}2lvmF Gc'Ac\B!>d O(R;ƢFY1Y2s:(U+NE1ˀBbk%}yv! " (J+T\3w*nLB8 $@6AJp2r} v̺"sa؎5vgTg}3rևS? &Ÿ/3V#5"?hăF\) j@1BL$9!3(CAX'#Ylb&_J6rƴg% d(gBNe=i!Q4)2~銜 JzOkqɶz:֋zq3ѣ+^ȜLRpAPH+$ Av&&ǘwEva;}@ #\ |8IgF9nw~LяR1:~ WvlV\cead-rRʢ]sd8A 4'rCo!zBuG!ݾ3O*) h(,ByVl Dp`hxo%WnSZPss3%NB.Jڿ@ePΠ ՂvV팜 to"ze%-+RpvŦ.iw!I +^ˍ%N:&7 g! E $}kE.x&sYtpCKp*ΗxTsdieScȈZ.,ܠDd@0$> }<6 Ey>yH=aɡ̥eE6hUE%9p!gyLQ-ڠ6LtȈ'i`X4-98{C03o-uIUJxm}S h dG,Jz(*a:OK!gYoIJE桔j5lq&yzcC.~+-QZ;R1wxqW}r1%gEaQrtLݜ^82˷׮B26' 8fxSm#LbWyݝV}jBGܢ+@K2TGcvtx:_Q=ExqjͿfK]>2ϔDLd(k 7]Yh9;csM{ 8;8rx\aF.-ϗćO$c+zW$=]Fɕ,)lQIfpp5EpbY٫`{?^W<+gkNj\BrdGHM_ֹ7*=RFniT#O5NboΞ?}s<}rϟ^xwON|3pV`?? Axtmw Vnе]>u;k|Θƙńϗ ӳ~)bjxF9Cȴ3r/o*&^lSbJ0~=ŋԨ+^u9h免neeܴ%>G-r) G-*5 v5F`a) $em===ϝ4j~Cx>CܤOX dP&֮:n{=pFyUSp8#'=hRIUJ(1/sWmEBvpb9D2-&0=|13;O|vByrΕBu<Ȫ/lG-;kyW_XJ̫e)ϰBܗb\2Y2{Y?욕Jw>‚-9zȃ%.[kI\ԍC[gn evVCddJtރ, 1i8ьMSn -Yz+¥G_U P))ۚI1[z[rܫ/|?:Bxw$͚e ASXt/渠7-Tbt%;f\-UcTI1%Չxpxrh6p۴->ɟO7غa08DKs!}4g8he\tVSz!Iv%`,Fa+ү__Ŷ.\םb,^)>7rV4F%߽:2񇣉Jwd<9f}HDTg|РԴIn>³X8kxyymGԶ!}al*HM]Txw{R66*uz/8VOh!U4/<[gi#=%Uq9>x>Fx8%I{\ ~񇇢Z/pthȣPzs{N/Vȼ! OA, U?[Kpb[gke웍6 +ULKCH:Cz棗g׹ m|s&"y_qÆuy߭bw~YQŚ'G{,^[Wl|h\|<~ñp#?|;twVp]@)JAW7ݶ/H‹i`-.nBy;>X|AMvߺȁЁſ Eů 6ߑuK~N6^Hup_KG5m9d-WY~|N|L؛O\Wg; hꤺ408mq77 E<{f.<=y?\ ?ܣ0H] tҕX;DWՀ@OW入F{zjbwijݡˮ@*WDWđ.pugVZ (]@WW=Y~7DÎ(lfQb0]ΰu{g2#"L>ru`z66]86wl4aGo_p']tqh燠86,tx~xÓ#9O^ʃ^;EE.Ż0;7QڊnK܋uiwYK]pCd* 'TG~"9Gys\~4eį/\5|nb5"G_;dz8X؛j[]>+{2L/d}oƓ7͡21!>d}WЮ=՗ild'Ѣj#77 bwR#kM4@BJJZ`f95c*$B;mtf4cw[D34fuCkc˅iQ{)jG@%-k}Չ"Фcse Xi!\Et͹0=.EB1 j1=$eR ߙՠQ"pK8\Hoɫvu*rM 1Px@֚,99#A)pH{;oNBWL˝փ[RIIM'x\W3 p[âV:j7oj-b'ԈSl*%ń>9|Jh^LBʜ`8ZO[5/*"Rut IEo{ksS"0f9xDN5.YkRCv4,,x!l;a[gR@x&UFa ,b X4v*J8( i RԦGj(JV**>: Pl XB}R_Xtĕd B2zC&ݣk`+3\"73 zKXtZ̈KSΊD|1&PTam!Н'(JLG ؗJZ&(TYǨd$]I*2:Dd4X 0_<+XX茸aZ R|$ L&rZed^W`q̎۸dd:m4/ P}%HhQ2*@vdm^ ,\]UBN5~d}gy)bNLcq\]-ZaATD4XƷۻšͱ5wy4$mbOet2zbM=!%DC!e2X xC$*z.x W;8L]:| $$(L)ڽ,9;k q[4GڙJ2fhNJ@A-D +P,RBPS`y2f h[6σCsFXA"cP:f7c4XdЙ$ fjbkM pyP*ZGxwG^\ky8 1n},,)fle'V"Pk!Z$Ot#, MLkdrgV޶ kneLU{Y"IIX 4>LidG `fL ImUFjhI>O h#[1@wLjv\ݶ3I22P*nQlFPi6R`7  ~;iRKC\kj]/ 1r7ofX"` d=߁y!x˵٤Ze ᚟4r)A Y 0p sx HXࣆ=xt z{YǠ\3ic+юjZ IA6 zn6< c͚ YS{/Ѻ6LZĤBF2B1{q$<{.CiLIAp֫LU|8mP jx:kxaALBJ3 _Ln iq#f0pzOZ I'a9P-t O atl\\;K`XrAv&f jr$W!8 4M]X4wR@Ւ[ Y\ʸn  !B`FY}  R DžRǣ~|7^CJ\.U$,-Z /Yͯ^>teKu %\=Ȳ[r7 o.0X=Oz'q{쪀1$ޥ@{gNZ'hH[_N /NNNNNNNNNNNNNNNg\{4W~P>`wa1$AkӲH;d/-jTheU@b'; N v@b'; N v@b'; N v@b'; N:\Ђl'Z2 "O 3 @ctA&; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@uA %'@ X@@) ;FBdN v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'h@ %NiBN tB\86ԝ@zN184|]o9UZc?os훳|u]-u)Ydw+\ſ~q2DeYY~T!Qjxۓ,QmOl{b۞Ķ'=mOl{b۞Ķ'=mOl{b۞Ķ'=mOl{b۞Ķ'=mOl{b۞Ķ'=mOl{b۞4áRNx;~7˺qryؼ\ kYKMǥʒqia"$J.1D\!:OE6$\Kr5FZYKND2rZP+1uBJr/izn*pLV+޸v);zӆ!/2!BCv[c^(&$WFGEVv&pITR91d:Dں_&wxkו3 }v6&} =ËMʔ;z&?O}3Y-7M+shwl&e&DnCh;hwXVF4;MFk*2.H=eZ%`oB\'F\ec\m }$#W 2hԩRvŰ\G9Br+5d^1 mRr5BrVy/ ɕZOFגIvf)v5J;I)D`O'BWuKH֋ )y(*8!$W)d qBZAt\\EgQ A`ҩ]!m P\}5r%lz9k5|n`koѺ :-=J\ImI +F=9MLY]͛X9=fwXy+ˢQUfTnky7JXjK&F\H;Ԡ^JNGbkb$W-UZ\!,W#+v\\++Ֆ\!uRz~D !QOg4F2+*lH=*Dm \!T iY7jrWV+֊NtVQ+ʁڏ2FFԁM~ U/`퇮]ދȁGJ[+rut;;hQg&Ό tS[ !$ _K22Qi 4PYG(8 d q-itrBr5F2Ro Ւ\!M^ұ\QFR+tjv,BALmW+$$Wz2VThpz C?(^Z TM{ȕf:*xBri2r^S+*uJ#,әX? r&L x4=;mzuHѧjy'BnK 쐞՚)/zu^Tҫ 뱝2&b_Ɯ*- nTJ zi)K )A:i\!,W#+ͬ$W<B\F\jreR2ŝJWY*r&a1Hi9\Ap"WH;a R rQ+ȭ#WkDWHCrQ\PT"#W"WHt1Ut5,+~-~ imrK_\LsFp ת~v7(Cbɠ!Wئ.J']Eډ3cgMZ}봽 TdTq45ܬ#TiUL`Z~+č\)J^RkʕJOHщwiM5I/J' a %BH'\+,BZ%S+Ԝ QlPSJlrL2! (7 c+^"pd p #WHklr1,W+\!]d "NԎjrR C #W d@Z|t޳\PbD!B`Ogt,>K/Z)]!+{`۩'o›gh~11!Wئ.Rb&_ZF'G8I{Jv{Ԓ<%{S's:1ݙ2 ) d=dq"W@L]R+1ʕǪ%d#WK(҆KH9\3 "T iK]LFyBrXĥ ɿDJc6 tr z`'U?\9h~z(mb|,W6"vF'6oŴ՜FrsYȫW˺N[-=21J'X)ER)- P/nQK RZP0mcu [ *,l7 i"µ|:{fosoü}oB;nr8z\_T/nѵ, \L*Ϝÿhۍn|^y} P_=wFl¾?˪?-ĆoGy|~{]=Içxwg7YX]w=GO͟ooNu8ğ/I+ۿ݋\]dYxlM%]ݘ2WUYhd:BiSE ۈ ryq ~]矰ww @uUٲnTd:sx.ͤ7veU_7&(!HZYoV*oD!}!R&72/}m>@AE+m>Zo3<*ڿ{گ^/$`< m ,חⱻ]?!ߜsV/ngq {+>|ug˛>5yz|7<(뺄~-Ru^!WlOlnX^7?^$wݼRղ4J겎FR# κtBhh۔6l߼'y?rkw!۝MuMX N$7bŶpBo<A&/ ZmF0Yd|߮p;qbg.f |Z>fAͷRqG71ͧ+䣅Nhr)J:gL}7w 3MK;ygRQL)93Om3թC /NnLrr>s3]y.o4p9\/׮;&|#KW#2UJoSi,\ e)LQ(b-d.](WYX9*]lVy'ô" uAVRb}nyyyj|QOK|Tbj:6\VmmxX~Xt!DwGȶxDw) #:[ OE竓EQ!~j?Zw3VA{/zHkH򃧊.]Vf:76VSTdBʂ :ˣQ 8z$a,} {%S#o!caP,_^*6*涨Z+[6r҇quQXFkP߂BZ') ]9gt(6n 7hrTѩܨ:!O\;+ump*2Z-?~Jb珯juqxxfgu]{,,ޜo^j*TBz(S.5NBFJUR a#eJNzQ<:Jbe^k:2yl HPԹ^8MՔ"?|2Y_}Z.77y9.N SۧMv{ugsA*҉&ߥ|~ܶO=T>S3ygeNHy~X EzUX23MgD6Y#JWZsGSC^mZHѦɱ\6Uȥ**][oG+vy$޳:FatbL\OpHHT 5꯺9eyQRؓKIyn] $I :ik*OZcO [i O q )T@$B[c@1IAI-MhךZHu .}Ƽ?>ݞa 76octԁnqK>N AyĻnT9 :f,*NtAmlJƒ-܇ hž0Cl[=e5t] 6Wݨrkq?aAhzwSxQU+NxUK,O!T(@9ttJY/'B.k Nr@X tW1QtbUu/Ut16`uh_n譍^iiPwt=V%QO/ Th/KLGqqzl,څƵq:1a<#l|^8/弸g-qH.6 ]H 'k{TxmZKTb0''U3WHŸ<a;eŗl͗lƗ VFZa<8C% JsS:i*& xF܈ԕ)ŹTW5s| k^G\SMϟ\xkwy$#P8e{W{^),8ZqތԻxY#<~yٹy2AE6Au]@R63}rȏ}p ~݅,J8lEtW#z#eIEKi!< IK\i?2ZlTvn> Sz[Z[o(XUx|K;a0eΨټ̞i izW?ߘ{z~*n

    681,iNofE#{ӏ6 1|ɩնuZd4`y6Oܦ~]񿧳 ϋޛ :0,|vsqxN[w }ehf֊Cr|zT?232rrh!@)?__ qI.٣+EěfEdzÛӸ ղIJN'7$!e5뾛CȰ*PZ5F!m#^'fC 59PJz6oNqs]jPJCŜ~hC7)Sji;vbmO< z'LǓQYInd#ˍnRy-PMا?߯eg^?9SxCa/VXB{Y? Мn-[!s[+k67za8OOG֬;wrN_{J ! squ\H`t5^Z@%ûsCg>LIerޟeG>Q֮}C_?LsP3`|N#ab:^ל?L#s,G߇nr/~}NQ_q9q]2?ϡ&Y{mc`Wf̥Z0+-l-6~rݦ+˔gWh}H5%Cs^p=]2"Q3Zoul8k @|wO7X$?W. I7tPsܗ𚰛6-°\M̴n"Ֆ-fM|])XB]|hznѥ.~24.Q!M ŭ a(j#Q : $AETlow-xڲW%( 8s\h!|^nKiv=IҀĈD,;8;We ]ժ^˭> [,") [ZK 7~jڏ[#}#,Nqc6UAG$ U^WQY&0<磩k^=v d){vV3/G'$,מ$\Rit! =»cYkc6PczjU, J9ᔕĠx* qKMY헿zxVg}c>I46}l`yNbкԳ u7X1^}*Hfx;w_6Pxp?VrI9"zU^#zMUT4SxrСԙW:*[^e+ЫN8%qB/2))`26hdM*VEbabPcA9,"RHV'BQYf /&ΎM[]vؖdJ4iwg hy4) sBPD2^$<-і}@"(oIBtSh 1*Okb|aC-&MjnU[VUS_Ub37&}kmDՙWMםMޛVTLa&W+B"BdNYY&j%I+ LAQA'OQϨN!$bA࢈>O,wJI8((%tR=c1qJ1_XL3//6DYd#\bd ꨝ+s?EVpvmc_H {Dy#ne԰5%\ĨiffCBda(2mw0 &Jc(6)MyЂg CE̻kI3rO餚R7yO z6uӒ}"/Y;M㉒+$1`u"å#B& V` qw~~-v\=Esbzw6I,?"yEC\Xُ2ay;<*@j&hp_?Al~w^˄]CՃV= .'A9 ؑO?>LZ{`i>HJ"42dL zյp`Me>w̾gggϛ$@,H='Ej|Sϸ2k_ήx5bL$)1 JfpHV,c/Y`ސǍ`V[:`vf<QE%].(nH` MMUjTH&a(06/h_F$amm. ZSeD.4 pxeR11i5g0 cVzgT@ӭ ^F"V?dJXR`UB.-t9^Ŝ>-rEw΢ɵgTcF5iQ=Q.(6J[n]`)Wx.D-9`a/%Ahô@‰lMHRhX X1KLTkyXiurd/a IұVL)# RPN)JC 7+ hja=? c۟jiZ+61d)X͘&F͖ >r 7I^s%TxuS & .q35$+4n' <礖Xq,}JR ͛AG~/ٻy 髃iHL>w ;m%߇ ? Ec7P`vϩhWiZqXLVŁbr:Fڌz1L cQm GhSn:M'b0 ұ-{" .KxK@@ o$W x6 GP͐|1ocS,Зʝ9]nL~H}ߴW2Gs3-W tؚB1Rr0,L' ]) aq鋚q 8kTyWx}6\xs)5b, bNoQ>~T,ˑ]q37|˺iJғⴵ'!o餭ލݬ٘F b=:_zsx2q?זo䤪@EKHUQ?ǓMMS1>b\#sjzK x◺ᨾ]'`?}ݻGwG~8˛^Ê'YṼIAu7?ܣkU߼kn][W7NOnPB̒+cv n%@?zN_! Z^u;"qJQxiL4q"XE#tXAa~ҮԵqO~GPĮsYDj|6Rjh`[dM + Z!: vL+#)&ɤ0& Jk'T\ h(h/> Lr1} sا ڿe/½`:m/#%ly(X('>0>f" ڨ<QJO5h)uY#@ڢݘϷ8Ē)Mا&NN284ԫ1[Fx],ϼaZaLKħH 3:;-Kevʞʍ?϶1ocp'Wp"h#N+DN1'C/c[$cTǂ3 .2f "fgP|&wQD݁ fe-vZ0-/ɲeɖNQkH&$}o2J,E+O2w+KԞ(8pA6& lyh-M!*'1тW&qO`<2Y1/NiN ]"錜5ʌnf'I'/M)R+{B3Z*[ x/C/F'9xE *7jzCegm x:^r;ܲ Ip/gI`{܄F9ZuQ&jtT4F`U1$+jrHv\3Δ%zXz^$-*- ,٤agyr9k֝fa'naT59RHZ%D`L,g\[EG5DJ<"=q9iYRL/]^E<&'  Mgʚ,[MfVgVvRo'=!I$:a9:$x$IŹ"ID+$SJ!ONcHt$*VQ Qt 4=I82fLzJG1Yu0P0mXdۄo 0̻Z>9ୖ;]ZgJF*ZH?cT_3T:la_l:H„Gr^R$0d_B]dYRcfD TXg\:N}16l6A5RnOPFtJXP@Q-9Y9YXD3g`)HkBU1mcYKI 1PRD༷TDY&1ZСxSX w :Vk;#"L;,ʱ({HL &i'XV ^iǍ WKv)t\|5!UQK?,>Uɕ2H ,!,:XtZ*Ek:  KN>cj>.N~aye[3*Ҧt(#$Q2u6!d|~Qqw?ů4ͰDNƣbi"t \D `}*c-dV*!(J ϣʱ1ed-*uo!tuEᜊhw5ʲH-%|Tl@<9IqzOS4PfչB5+[$4շ'{#FEVŮ rAӋͫj|O|soZ.tRUleɖDu Zoҹi.6VoyЬilq Oh'Kx~|yFܽuhV “%.?&LzǾVT[e[gQ‰Yx=r۵ь+OKfP7֦|e#۽7zE׻W@dg" \!| Wj)l(Bj9vB*E]D2 \!F \` \!Lo;\!Bp2I.@PW@\Ip*KJmLW2kNiE0~gvߏZ%G2LWN=3<|rF$BrBjWbo)\!pJSD_RÅV.bś*h `%^|M7~4h,ߡ%1H=0hoo݌ c4lFrFj_D*{E4Z+Cpw\3F0RɶRiHW/OPBBrj%WH%WR0w+$ M:+ֈ+R^ \)+ X]! \!B*Wv{z9p%fA X 3pJ+pjpT|pedA$XɵtW 5b T%•]Ү`+UDj "+=\ 2Z8W'DMd&6LYv*k/F+M#z2co>*yP0 w}ٻ6W#@uo ,v`>,Ei=9VLZISMeHImKH6OǹVwWl۰ɯ8:;?X6ůW;=OaGz ~7$ nxJ# ~7\³~oTcۦŮ/zm ߚPtՃ*Cv:D3Lc֋\ 2S"e8?=Y+kc#R*Gc<.H[W1B㸄V8ex>Uҟ.z\zuިo6OVy}(faruMK;!ඊ&ڬe))b[`n*)Жl%!^I0rSd-CܽHxbuqVsu!r95fFbcqn9z_=y:gY_&wLA,b_C޵ߨ7թxQ6yLքoz;C/ kŚ{s]$zޗE%~ȱr SZ" kWְfL.xq{vR@p?8%-?nڙä RZI•b_g`림y8ӢF`kYJ򢮬ӏqAĊ0U}8w6/ՋNZ+SR/;eӨ OZYW}}eH_FX]rD%eE5dh0䍊U wYYqN C|e'~naٝN.ܻh8=Z`/foSݾL--zS @iI?O-S D#7tPQ+z+4Omig&t%oP?'n:\2BW-ۧNWgx%f {CW.ɾUG骣LW? ]YZUg7 ]-ӓWW%gy>+z'H }=jOtЕxd,:\ٛ3{+u5~~o pڛ3~sW%3]}/tEj7דztuGp3eR?-uEjwZmLW;k :'}6/~?Z,[?]\|\I&|қ]ܔ}޻}"6_D[dنV9!מ/ㆪƕxF*|[X>! ܲAya>0D>!/j{fyd|r3@=z#).^?[/#9xVR;\7yomƾNF-\IJJ[4#LyWΎۧ H #>E]CZӫZtՒ.ԏknR-AoL8B4l%d(R X}]:/bU LH13Xj6ڸRѹn Isؘ{)cFE'u䤱ͥM s'5QǦDFjsDOIi/vNX+"T{哧Pk8 g1\#nbK1p &YD z宁ܗp I5KKl]3H9) d8AI$JR-$Bbib nlZ[B4Cc6mm8eZIPtB%C)Fe]q["wj!Z&L!9$Bc ~bьhѡtԔR0 ː@p+2Kk޵,lMVpV+ .XD!=+Ј *=I}<}WO1χ1Ά1Җ-9uH )Smeu^Ob6ΛgU&jjbu9 U'HVWZ\ V3/?'@1;bequ~akᐖUBZ킙)PR@*,R!!LjTm#mSq%"`Odgsf/)ÆUS*foƮL #1K"etf,)j%D:*D{RQ]J0UHP/>q*XoS V\BmQx݊ *9m[_-uZqx%(v(X56My 0ybM KDh#`sD%lL+Ȇj)4 _'3T[\}`t %`q(JƬE5,)ȡ3" + ջZ$Z0*!EXl[Mnpչ/z ƊՐ(@&:x0f L[ 9G:(I?i*8AijTUpk=8(&`=y=P B@DEP!ՠwSA02PƺOA2Yo192eP@dBE@ &͊U:}n:oMI;$$X,xΚ`0rGP!!.A "URzC':3X%@ i㺣PVġ~ζ7+@TR|F s&%Ji7 ] Bɬ@Q 2|_ (ĩ$4}1 9'0#1YUDI)Ts*#R)~T@B¾$ZDŽ@yR"Z(s}=mb A$z}S(:h_RmF׊XkuڄB}΄-aOƊ q }ƬI@qbh1FPTP]!/ 1ƐJuu2x7-Yw[.0Asm`-d1^-*4ʉ PY9Hrt9|dZ. p1-(v5M5͗ -tB\X+hjIR|"2LԴ":#>8Ԋ.L7}yL\t0_SdPQx3$nCe6s*Yȩ ՏUE3vmE7Yre5 D!v "}dt럟. X󨓹hO|'X קXk#td(cѣQ.Ŭ}(T`w}|$#tKLC_ISДA5Z%LňV T׎a:Y@'^/B>Jڝ+]`fՌ&x,,7-Ġ^Ý-@'>)Yp,FJ3% Di9DHty Z;XtDb.jt Xt*üOTTh*!ΒPBЌD%5Z/h 鬖9k4Q9tš@ r›Ak*RފE8n`!-VH =SRLCbzihmv n9\t>;Gq\ iڳqpI4}`0u`f-=*58)]Iۦ*'2 cԦ¬1~5@<JV .^DX|&iVZԡs8ݐh!/Qk M!TsC<ZmqnD;5)Qr -0%4!J,A7oͰll+OV"bIJBi$蓫v7(NznU腽Phۢ#$KbGf$di"1ts֎Ϩ,U?9}ϟ}2@ =|sf5 iOo^l8xӯ.z 98y~zTE.VCz^ĸJ=>?:j8Bロ>^{z{~~r{CDaew+"$q /F_Wozۢj;ja<W]gs+.L>7_ fR}2Hepp=L@dOpB྿O@9e&(""(""(""(""(""(""(""(""(""(""(""!Z4ƭXv{}\PPTJ/rx,̔l9ktns 4u` tt1[\EZŻ.Whqc$W#WVlT ̵/z$W`㝏ʝQIGS0kBsmoEk}w(':F2Yz$W Uo }S$W`-rVJc+ˌGrFsW}+W2.Whh(W9kd1d+0Kr-uB+=-\y.4=+~to ͕/ruB+EWlz~O1po_Y2WXrҾRtJnC8cv ~-z#Wh}+ֺZ91ʕ5g~ܙϜXN[Qoav|ڊac{  cprO, 4.¹A?]s"e1yH0"yJ \yޏ'N:/C3J1]Rb6'v^P{IҾPe@ZQYWrRTMJ}`Qo!X~j؀:14N85_ )v֒’VU~W L_c71Hk1OiNō|Js)WH)c76<ԜRI-+vR>e!Vֵ(jM7U D=wf⿷:FjbV-Q[xL`B҂_fi7K `.-Z{ZZ8¥)4ӢGrk{#Wx̓) +Rt+%<#FF\כ sp3uB+1ʕ_'|o U/rZi[1ʕھdre7Bs\v]JGbQtViœvBk}w񁕞R1ʕS>]=z2,w~2Vv-ՋȕS  Fg_>\Vu]JOKlzq”Sig{^4U r%WYٵb$Wmz%}+0X~jBk\]{2Hr2re-c%P\)v&^r@ݣN0g&ܶx1*߬;˛-A2^mR9z.bΒהkTVx}ؐWx%|1~a>4ٞb؇65]jYd2ϐ>|_\ͤVh6<e-~=Luo[2l}_ ?LxÝ3 16#(eo|\/˯܂$ǫ-V9jg欆fM#V[U-şvra_M{ZEY=?r7U|5بCRVMNdF-OlKޖ6V)/违MbAo__F_ߺ>l >*3ޞZ^eټ杷j~1_GJ%"F=¢vuĖms=Ul3S?/GF[˹ڂʮ 7blLNOKr!LHR+<\SM/j!'Ń̹ kbX9(ҊhǣGyʜ\sL32P׍٩+eyo^x9Fsd4Lv>xzi>wغ+awۏ^0n_X}?wtWն~n_ކσQdaIa7vSo7y8񾽾C[Ƿ6E*._g)F[6q+ڀKR|%GXjU/ 9!*Χ\jtnzĀ rmF=L񟟰(9Pˁf}-h[%fQ%TüKl`纽Ч-B }S k)RjžIŜu'Q r(6dJ F㊔"+E Jw ^89VjƗaov'to-*NG7^Һ c& O&Q} Q3:Zom4Ч&=<4 fL.E? M5 1uUCօ; S›~X L8tdVe0b\3i ]])X$ ^f?fqnv:*ݕ5E]R0%s*,3e,bYM.V͵qtJ_ t*1CJޜV t$"#RTUsQ*P`VZhD\ە=WWͶԒBQ3rAHiX`R&aRQ5c"*.YLqᘗnѿI7ev1u/'G]++tڭo|2Y񼘎ʪ&_=S!ô &VU`fbd2NKEȘ:S٬q|6p+"`_W 4+w.5{{8C11H=/Vɵ)Eoɝxk+g\[w*cV ƚ4~w<;"Fq8 i/iVl"oIi ސ@M^i~,_THm=g67[(t@:H}ɿQ,{P]wnx>n,xM5GnyM5ZwgSͣl3>mkZzkVI\kƈp %J,ĭJagm)pE})1J.*C %ke:|jp`|>&X`UJ&H\NQjJm+(0D%`p Um*c_x1F~L+x ϓL0o+`5fe1 ; m|r~ߠF]k`Y\j 4QƍƏ1 $+nm<8_۲= o$,?㖍 ?/fU!NB ZL?=j[oZo.(pony i(+o7Fi|N~iiEcxSۍţ \mfy)tsmpeW6bﺽ_WyPΕwdrGyw^W!}a5[Y_~Wg/S1]ǣ/t&Uݷ !M%KH҇.U\,\\ރ4/iV)TOփ-SІ9805:*: Jq($4c7y lI[j|R<@|6MoFw԰ԨCgME5*sYJK@G[!u 4rP0]+d8l ,@pZ#ScD, )A9fQbT1[@76l̾4U{5ZMN gIMts 6! * 29~"7zջk;GOۺ~7? O-L}Sf8*>r_]k[>{:1 v? ]j3uOj4:g*\r e%{3YI$x$BzVRrivYINVC'mmOJ GpϫJyX)X \ic4YbW6y%|SּFpٌt[ֵC@Y% fQ(P}%=⮷wu>]ϓ>]Eş Yl;Ou<-ř:Ҷյ 孺nBrO|Jig¼]q?m-qrˑW APP:W9-$'!to^Hs:[FMprUYʆYB&i٢,́}B˜)3)h%9С݌Yz` M̧ą`(ST[o:&Y<󌎡jy 5qD kձIFX0+T:Vƛ'_R3~;jӗSGal\ X Ze)R9!&P|2cs:dVDkyt{b1NFc't@,5Ahm`ZVR{> c,,IFfXeXҙW.ɕE xVlSc( В0/0/4(y$Oc`\,U]PuPRB @t+Nd$sZq*XE?ƣriH/95zK;gWZӎIxjuXwRlB|u@%:N1]db݉Ǩ\e/@_Z78vbRORɘڅLC9?ZL/6ZpX6YtZmqT6~ߩ\?b7n7*|K\Rp˵f2 O>( Qcxnr3^-&* (y22)&:p ƓGFW&K CZU2=셶wKEiFRq<!]""hHGZ-+APM/^D^?1L$> C?X"}R?CbtǍmsCMgΜJZ]uh}}8Ի,`? gWMXލg/RX(=>uأ9QN r W2<] D*Wf| ˑu]j#7*^pҏd{[4lZ&+:MVhkϘԢL<:YrϻW~Xs~Ir7,3`邒ыzbV|AXug̜x[>E~B7-rk>~%tpv| Z6O(c9ћyRy8!4C1"m!7O3uo]dGxF|F97v><('w[c&~zh$r>M$ׯ8p77,RsMu7@*l6Pt9dwD 4]e\x_:;)L@ͅG)UY9[ὨOy'(liaNjоya/;.,"D+JΡ+9ef'}ш@@+]NRt'o2`~,`^3 Z.OO,H|}D-m O6JEeF&]g<@R" )h:wh'>OJ $[z4wWY$u컋Y7tNQS"+-!ۤͥYۯMtghWcfX轟v4z_4g~];j~7O=wtf|/7[6r+kY{ug=d!x2'<5𮧱'g`]ۈ+8a]4"Юr$Prx8dCA-v$di2GJJ$3[0:~K"rpȭV{x.rd$<*4‚K2[g|.79\5ӻ gҼPv혽4EDZh-`8ۂWYe'ׯ-(S1`N,TrI rBTWeM ,Z:oNr頫tcG#HȿYol!&e4 'h~:P\(: kڱ8 [e&/bSI n C@QK  pIx%$ęv%BId}jKcT8-)JFQ"_2Ge0*)YpnH#'E(I/E OjI8+K@n٪dGYYr6S)kY^$ Zq˔,#`ȚZI;HeD7_ewU9N*R.Eg0(w C f9ol FkSvə[%19>R |*)i`H?U1ckG퉥řS|ӱwJJ`,]R:/o]L ͘nXQ띗\{%9:.sݙIIo]MFSfUx'QV}eQqKP,..8Kz24݄LY97i%h\'{;Vk9t||='0hケ z'Zh&yxG:sc?)f_hulyL]./ޭ>x[bk1O'5v&Gfmڗl|*Jv+!}.7_&tM5?1Z>{qgW7zq}͓'4W^5rU7TxmIt6|yv^˩/ivΖ;:礖mN˯gɇwlcy^ɿ_7߽z?^ywb%zHI+*8HXYLT(h|z u,}혒SL)pD5HO=p}zӄCXT6YjmB]s8"bվA#MP%@UXUNgw:3M_ρm7;ϣuenմlg?v8uv\uA`q%%ĦJ,*3wTں\gA˃=B;ٹeïX[aK MU-"R?(e*JdePR@I7=G(kOU|$%4JEPW!D|4ƺ6RVOL+oiL9 #W鞲Sw6nuW%x}f.`8yG)htbUTCS#mþb^*H*Uh4I!]4aKy@mR !xshƒ6鱫 ᢕɢSmH^Y$E$S134@ؘ1(dB4qwAɗ)čkf"f.]'_$/Lpfߥ삏1# 14Gٗܲ{s"dS/ 3>6{-_[CMz_\Iͮ1 ꁊ';k}ZHo)_|?Lڤ;gKSm]o~rit{y_OQƥkQV&(h\;gt9v)aU]mlYMd_P~pt[hD[hJu4ʀ &ҺjQjAM}:.յ%Xz2.2%K WBJB0U˪ @7*`Q +)UU7<78-܋1%fv 4E0z?yR,e(@ J_sxG傖̀F5NxO SkV8mx`+sCO%jz{ݲ[%~yDzb0;=sNq en_mMV73 VW!]ȨM&҄MiԱ2>9TlD -r֥ JF]cSUJf1%L$qr eit)eJ²aH M|*p0x( ,  ҺDťx`/l|yQ^0}ӓO&&Y|ʁef烃PRZSU}QQlB$@jj=l?M|$+V8cX':-F'oñԩQ{09_'vu6s4iR5tj?4DC`-*O^먜ԕl%mit2b@T~)Z~l M9(Y Dg$ǗE]I6% EUsP\Y|mPUG8"f%" LD8>w$G!E5|!TT'WW {kpǵM A)cDh 4ցu#"_YLu͑F~4%y,p:)PrNďwu\\\\`VP.'.N\%d 2ѐ* q@kM4):ʲ\ <^oȭ5x5(E[,V?iE26x,|\g\.XR~AM8n 0\`W"\p%jky0\ZcǕ&pu0()J"8@6b EcJTNA*8 3•hJlRz0Wfˮ7Gy~'=㪟\OU/̫ટJqpЮ畾nrCٗW\>^vބoAו: 3Ed]__yw;g*gkW|*-ZbiujӐyKwiXD '^3/, ;AnZejUzXX3/8ڭgo{4{VF'ܪLJ |ZeozWKV z21}b;2wmN,BY0:U**Nݩ;3Sw֝#Ҥ 4*4S'cg:Q:QtݳB cLU`lZP6ֈfQe߳&r EL 8bKB2•^r \ZJTZ5qŏXW"WgYԢ;D^`"(gmFb +ky*j{bAMbW6Xm0#\5:\Z=DsW.HEp%]>Dn&dFq%*qj?D\BJ; Dy W@WR WWeR7k~r S{U?42 { '\=A7tD}|?+V;D儫ĕ Y0m'+=}ձL.q`40͂ P6sE;vLJr;3•Je+K. :=aW!W|`:\cǕdj4*X0* X.lAVXe71넫.xWV{X.: X-Wv/'\F+M6^K-߻f!kb^e+\6 8qz_z485e+lW,!+V=Weʫn9Փ2jx Hy.( WiRo2z5&dSmgЖNMsI6Q+K%^DfTzW/N㩗Q,:yqwJg4 =f3r}>S ֎~&TT=M-Ԃ! d+CViW"C.nb5=pp!ٜ+l17"׺\pjRcǕ,5pu8bVa=3 ~-CJl+Vk5Wrl)'\= %0#\ DSkF JRW0X! D\p%j+Q&\"`|F Jj D-?5 +lW"7`PAQi݄g+e#Sg vNGO߳wKW~*pe{Nzh׃W)0-]1#%7nnei fif:3vLJ'L ΙpłqJ+QTz5qe|f+lw ~jU1puBɻwN~jǏ+VM:@\*&W ,)p%r)\}e)e&\ l,!7N^P.ڏWc+m0`Pr|6W`SrA*(kW,u>S"d }Փ WWnˮwGJ z[5{O2SGfpЮ08LP6:\wǕD;qhWCf± G@HF l\ߧ|j-KPuܹ)){F/N1褗M^ة O6exz\͸k\BMyU^fJ&3ͰƫqM-`q{D/AO-J=̈́Ԃv_1=7Eן\)Pb?U2ìgqդێEY;^,^|7>_9o/į맫pЪ%kү:xci^r]k}EǀUgS~"_|B3S1O6|9"eSuD6g7EONɟKM7\n#ٟ^?-G)%Cտ lUrcUW< Nk X# $ .VcZU/cZN鼰~=v{=، ^ Mbdﮂ[6d*OhLIH%6ۘEhuǰf |a6qJ:9]kdǂ Dv,XKdj}d*Sdw!~c~'ްdqqy:e3ntDMcdbxt}n}#oz`o,s&Nyi]E74^V 6n#$mIv]ƛtqPC^9 fF3.rM6^F~*&:g|ѾOQƥqbiQ}(GpPc(ȕ;Rʻ7)AQ/:{Lˇ*?Y:srZ)Ϫ6O^vaA4n )s{9'ۂ7){zg`X7:ƂQg}̳knutpfo]CP`1򨐋jɍ~'t8RMڃ$VȄj̷>mRZ0o;=Dymt}#8e 'd:DydB"0uy>B:rkXKm-yj-i0`kI] @j5cwXe@tѠLF."7d+V;D儫V6m$○nyQ?dcVe/x_6)^%YI,kITb4n]`ڢ_ 8E+i΢qHWDWXP'*Еe n/vtu6t ^_`Υ>a2URp>1]B+͉ZƔBX+ձ'(&5i ]YfteQ6-wU=tde7]"%K&/X"如r.[DQi+i M[T4-F&;~4 %i]`BMk -ZINWʷHW\%X @d9Z!teQr+!Etekp o ]Z;]YttJJTrWXP*p n ]ZCtʢЕRP"u{A ״&nP%]AJb[DW(we*7޻(YzteM+jO0hӕEXGWgCWք>rpJB+r(iü+SLGWǪh[ΧMެVFp5͡˘)D4qy9 r94xD^},N9O#. o啓`L4z.E&Bvݏ: 7:g-4Z%GؠPT04Tv!# iiڣRKJ/ƜQ>e;naW2-nSjKA[ZpiKj=U]R(R o1`M "mh ]YT,ZNW"]AbP[DWxW.m+ӦӕE)tGWoaҴɻJЕp[ʢպt(uKGWoF)w%œ,\њm1M+ ]Ic iS0hm+V6޻(uGWotֿEt1 ZUtp)k ]ZD PJ:zteڴ8yRpl Ug$ ΅(.z/0&䴹hiND)j8]-1Tk"WKդ-tEƄ7,ʆUO&\M:)#kl->E"mhw`>go<܌.w4i!:h!4DYTos%C!B K (&62Q) Ğ8a8 BE8!F;1&+j%|ަyf) DkZ $z+e:3ɫ2& L-RO)uce+c2S ,W/ zߍggVE0*R84qHi3bD2#"pjI53* `ތnf0M@&b]x}L"U5~?Ï=jD_>'T(ZK`OЉ6pzs'Q]Ķ)`WЭȬE"3n*!$ $;HQJK,% cӐ4L9!LSf8x0]=4Y8N =8L3|Q|s٠4/۽xItUpLɟ~euJe%GzܑWfM~:ex&s'w/qOt6_د!}JVeow9Uv?I p"~TDoq28%[; "5ǔdVZ)fUdҗg311gS7b|-_[*  Z[U3~_+P- L`x&ôs[).dJB[;Cnn 0,; X4Ϡ.^G6]{W|yl k˂A/GՍux4L@ O /2+{aGTd*VjV)<<1kfnrհ9|IM$is#^1~XyI@7w vomkU5ؖc2w-FZ1z׫B.(5}`6FhR)UI _-c<㻳K)nckeAZ='y!,hRugCispJ쭦[ "c6di\聖n/p}YywK`p^r3`#dMnG_NxWy=J(lb:dnjhZ޷zN0,7󦍲/\TqD0'3sᑖR< hL}k@.L+R8_[Rt70BF*F_ONt~4PBA!Kє(6]:hA}Qh: "Nw'pOy-,Pn)YE7_^YʾKP9ɠUU%I**tit7tba<g.XRqkC_b~-xͳ ?g_y">Fz}9O7gA^Ŵ[ O^0qm$5Z|+X.=k"%˚RKWw1'h}bhfD uꈐb%L ^ ~x&xpw\ع#X} IPEkĻnĉs>' K .pknEn Vb95zjGfLLϬ~6Ҝ X A$ qL 3~qsz*CV fPD!|ۗ8S_N%Y!ss"B;*7>c/ϺZTSrB武֗{c\n7ŌRo?- e*Hxv)bHžEH=d*UFW'v{PrD/Uض]XF!Gfd*YGcGcYEEgaxa\%K| Hp]q:suB[ȫdUG" .6tSg8b[3U3v !eOEBOSRWr"4jO| f̵C4 f|V/h!~\Ǖg5CHG:D"VNjTȚpȊ\;I{@Wy0*}I 3FI$t|'\DSD$: /z+'%mފ;oo;9˂C,!W_'.ŻҜi{}؊zbl^z*^ce?}Lֲ2[JT mb4N!PDJx "}O%4 Ť*}=" EBh&uQL~dg&xrfVE¼@n2^ӧ)uK:j~zܽ?"7aPjaAy-qG=j!dAIMi$;lj.m%)$6{mJ#WF9c׬V~2 9'IzA]̝famv<~g^A3ziՐ츬n[>_ht CۧY] ^ONu4F쬑Zvpnz^s:gg9/!WMkq!*t9 <>^t][o8+A=. yncyZ (%8c;ʯ_vb9l$%[RO7R##yB\7:ШhIfB 1!:YsX]tQٺJHR|T$z>j@QF"" IQǢlL.*arXo5KQ"?eEȔ7I!Te:a)Nf9rלJLx.V˷&!(m *C hfr.r!R5g~7j8Jnt줆k}Y$*d \L˜Q\N}5-zW>%rrj*gEK?@E!ԌSC.FHt`X3XYgzT KO{/N>5Jha|l~&õU-,K1RYafqR&}fr+1:^mgn22Ί) FR]㘗Q!(>t(w7K6e8f㆚hcmqJ2#2 9~O~p@sr ]bQ[dLMtNU u4O?gd҄Pl82+83e X8"TOb &Kqzd"aM蔾=KA^QCwV)j4)TAl8yK&DFp_C>=̞c!)Fv;Q4%1CrlH7]2Kq$ 0b!'&9rvɥ0V/Ƥ̀ɢ>-C2*.D-TpH8 "TYZTI~l0(-뷆DZU&FhIWmP/u0__/Ujk;TJU#\CJ.!~ϰꪃ7)kt@` Fa n#o^ RJzdons=1`VkL1)OTB1{c~`F?[oӎfDvN?-pwA~3218y?Y:F),>TA@օFFOkuد ]1ֳu⨰:`nHX=?P&*1~/5٨P9,loBƘʺoQw%]VxE pE&r9]٠ץȰFƐj 9D2jMDIڙDIf8dC[?ջdjd2r|zY$'8(γ´2Hr Y52*/G%P]^Ѫn *z^w52 fb P76Jӱ6[d6kkVo;9ǝ61H/e|ݬcx\PzOMhc I q@3-2kdS|ل;տ797݇i2˾rTk fS淌Cr@m1Gqj GxsR_ma~k徭AWa{bK%Nak/ÇaTmBJj:SfHQg1J ,M(48!2R*aR|O^EUvV$Adt{&j~FO2 .ωKCi(2yW3bz@:eCfխҏQ|'_#tHvGހk-KZ[mM <7񅡙=2fƎw6}ȮJOJčQaVav= j[|8{jַ/>/4`6]OL[>)yf0 Nupu=z0EYNщ a١>~V#c` h:*ϣs(ʍne23e\DƵFfOJq8@`!D`aqepXy}7"@_z9IWE_׻rd$n]ן`Q|:I'MlU(8cm#G0`xb.tD r,<wϲDv*Ho$f?KX,/K 'yvP <Lmh{ !(xY'E9f-qyg/ЉkcUp}P}l;EqqkqYPvɰTF1h}W1J74}]u )蕔I z=Ä(~wa0<}]_ րSݪ4oo DhtD-hRmoXJ@= wKKqԋ,$S#_FS8AE봚~O4^+S4~)-jb9/&\$^V^+BV0Q+3:dz,T#=B"E)[[f_[L1Wy8("$dpn,jae_9$.A޿m _ķsM)5ֳ!Oߴ3ҥ8.eqL{Pp;R`-<&ǁ{}p'AQmem(KDXl FI,Zv5uP%B[2&x,Y.{Ӝ5썴]=uBq*Z+:>fűO+[izZ<>4!raԨ1V bE wClAkDŽϖ~'D$^|Se58_OsIFCF3q&!ݔ 1qeP0$Eй`'>/uOoَo1r0mFKF| 8W3[m&˓ש]гf/^*;G2Z$=hV( NkDv|"F: Q 8-0v!<!?V Hu|=1!:TF)Jgu=hc~JfF?6xkO}a[ϷngYUh 84NkPI|l=.uw]n7d1,>]?uQ}k~y-/n6FpLXC1QVHL7!FVg%ΣADcGK EGQeO+qEH b0;'+B F8d\⻩dyZu78"ݻc]VLƼGJvLP'oT|iBqA$"+ i>>RȖrci(Mk˝wvj>RSH8kXE/Jl UϦĬPVZkcXl_r}VD$ލIz)=k 3SFȑh4H}Y Bn{cͲu9fjm7lʼƸ}׏C\HS=6 6jʗǧgт㘤! mA)\.?ح,ؼ,+X}.W]#D0NjuY DP$7ݎYлӾ KE Ca35Dh1xsHѴv?WZf`%Lz-F݆VR ЪHkG^PxgHdװ{^k :$ݑeq(Kk-Nz`.0c;3 J9t.3RYZd? p!Z9*]&As-aA qG0TZchW=0v9'@)v7vJ HɘX2Ӈt`Bɩ@!79J<$Rokmp~0/ԣ$RuR'}};*2 -x_̺Mp%\b:[}k6ì%#R^XW:skQw =h/4>KTĿeR<"FBHF_ qU?X?z-o+!KeBDŠb2LbLCa /dk旹8NN -0 =\@_*[s"b0&Ya˲ lk!}kn]'1=-ni@M m'ژnRV! hXFvFSf.Ԁ As[۵xѨvsK->CN\h(^&bz8qPy#=pjz6붉 ]`f^5ȉ]jł"q–אr_M 2(0j&RXr8H,p_GLx{{7eb\&at}<ɡW籑<6Fz#i=f4FA}0,wh( k,.\i"cZUxc S_&йr\!)z*,Sx*sOeyTgJ-^0CApKB;Wx⓵#PvcX5<dž| 0$°, ( 0-e?Ɨ]>)AF8ªRҵ#Pvda-ސtpt96&һ&2SP*5ø'%SirP؈nR=;;taG>N>Ij"c5j4rK1L2yuvsuͿG/m?J{ 68v/4ȗKoHK7;ȸԺ뛗G9jp/ilf`bJ0XOF6ho,4!i4+YkXťbJ+j=N5yphidҡ~Up*^2d#Hz00+$?UZT ܝjԃ(㩎<!]:A,(w\Z!VoYX jBKSh <\8ڛ^Q#Ywj0"m=0%ȂKYײxK)D&% l$@̩5`id{Jۜ RHB}٘=CrRٵFzm 9t4#$"BPcƓ^~.h4{.;CD}[Q!mtOrLPZ"킁DynJ=oh cx(@˂t^R<\T2{$Ut! ] }٠iw"{ɺd:5s% = n"=.)齑H$XtSc.nn^ ]뢒˚U9n|a1SJtצGeB*hJ:(E:/tKC:qxi \0oeHTt- 9 >;!^ 8RJk=ŋv;t>3N̓HRQ&{8MO-ԓ?fcO!;Vf_2\:ol| uv RCù7_rE*THKSA i(gr\94#SwR۰NRq0Z!$jryѤ䫥ҲԨRqN9#)Ѣ큱=xBx F_z7 >K{zb駭fVMlu`M[MsTOICL}Ndr ᄪ˪TheoOD4^'A㛢D4 9fJ kvG]c0{Əӟ?څ0ge1,хhϤT+8 :] .Ru/ HV,3b'GaX}Lgۈ6a#0X췣?ptƺ$+0wq63Bk|C,%cu.m"J;9OndBڸ)x)Qzmy(5UXjB - DHHJ f T1UG:A`,Lwﮭu{񏖮PSnos(u6T3pf2%,V,HDzhI +$wΑC^})HlDO!Iqώ-uMGgHZՎ._n*HX4 `O`EFŠ@<X1bh0% lY1VK]Pe )J鄓L)o5.i5. S(YV>Łr#8P||p\E);":S n+n-طbYݴ@l8 _8Z >DcqcGbhGOϾ3Oqb$ut[(@ѭ-phxɼ,MFwía9t7 *;nam T!X乿/p %C aOUE9ְm]1rK=CqZ7wJB+lB5gX`VAy&o'I yi,eZ&ŷt+/ ~˨prj6m#FI+[1C#GO~O*0;,]If>Qb<*5ѕד $4NfDr &cissߧB"Hw ֶ[j{{q: qg52rTRc̭Wl8o ~>t`^Pv8\4R)ą 'iPXRc0|{K3{=oR?JdQgY)p_I2YeLl&nUˉ27Xz1|<-B?Bpm& TGnj*$Ǻ{ߣ_'o']mosw*+ )w.qg(fR܏/ӥ(8+9$`UȖ",TQö;/ap,7fopѱ-8,w,gt|s1[DSŽLH3GR[D@#.(EQ`D4 f).Z>GHĠ} GƆRUAtE2`\ݎ"o,z\ҋrn+.`Z n|c, c6x,] WB$zl{5gfϪEd[ W%iH]-U]+7/ S¨ |!ΔES(p"R d(Wq}M|&GVYEٕP  )  0N*K&Xs,bPsgmhRHjO_gD>w_ yeV“#n o]SH8KũH2cNjvT"J ~rGa,5X$ǤAxJ֢ݮnYP PWNþsI/W F11Oh,EElhDKpr #!Y"i MjqBK~18P[pr?d8dLg?Nsl X d֞܁U5NKzى2ޘۄ%74XfCZ9_zW**HxPYGNDc*iJ ȕ>LcTl̋-;x:ƈ'sVe(I*":JNRdQ5 ќF MqK9;\ȞpaN>J"^ԦuwS]:^&?^Ov^DCѼD3Ww\mcB00>SivnS:vo)a4&p|tTWRT-1)hoF'S #=@L#H,{*+Qu TG'uBbKB#e+XǑT I~Dc|.|} FL5v0%[$$ZkC+ "t.DBE )1x|sP! /qi+ &(2;W&]E5[sY`js*bWd}C[l\r{^J5HB0fָp9p3q:_슏S;1sgV=2m>Vf+9-" N&l;`_r&&Rk'2{c}j P4Ac^3ONs-p$a6p=_ G0G Um|NsF|u҈Bp71&ռY$ݛ\k\x^5Aj&lWWDgqSM,S6;WKl Wz՗"ͧ9oE*X;꠳mԧ5;3M۝*i{•ut#1O6saenZ-W(1kZ]nSP'ވ'oGkOnOk|0d Pav>O 86+- n^JlDNjٺqrwc?gPaN=h ,܋ې%n>ier1 {;F̝ Tw {+-c|t~yO0;#N?%7?a@`<||& .`s3xhg}-EuA41g)j³ww9| $8w#wpd7$C`˻3wE$0;mOTJIB@Ub*!B>Z \#!:[TH{gngmNjBN5;<:<4ew$m'c\;>O0UiJqlNڹ|Z׮jfE\W Xqj_z2T݋e;1PR4&:T H;{(nn"YߊoG8o ^;y'<i8xMHx#SnIe[^S"UV 5bj9K0'рX>Qx=n}8,UWzZ+M麧fwZc=ÎpJA4;PF DxUYi[n-1'Z ˦C˘ɼг;9rzdv`Re2ޘJ#͕Ϳ>nS14MՉxqkRׁT $$ژ|Q4d.0)3mb]>`Ҭ#Ve %%ci.T*u &m25V%hc#bt99k# EWB5n Kpiz5nC16Mӽ;fAZÓ$W Tl!v&Hsіyiryg̔y)+t ;jHgUň#ɤeYz%{}b2:ʴe% ڍzf5zkZ^ E[Vzr dt5Vexh^,֫@9TuG7I+FiKifJ>k2: -_+]"UT EKF5un5}Y,hW{wOFinXO^˩!uYzbH;`O.|4W0RLq dFs6k7NucEOM)ݹr a͍dcqc z2fzB omZR[8oPܮ[$禎LST'g䭀 -|=nx4]7`?6,%V&8j>޵UPn&Y8Їah˸˸R[)~s5ެnجV$_ ,z{ O8!v7ts﫨]>Zvkbrl>ϗ ;ÚRQe(CM`R*\EI,ݱeT8(xU8<_[EIo"wĜj@L5u:6oQc까.WDb\pehE\p 0{2?QٟJݿa|EóGd< m.Z2Fwꇬt;$WӑNuv>χ_?}kbxW>ax6X&1"zҍQߍ <5Ͼ5"J tu/jWs#;yuU<1Y#?J1zy{2N#n;F*L9P$uH}'փEWav<+O[sMV]cO!%d$SJX+qU՞{JA 08\5!{>:0t1W؁~qIɅ++6dQ1k!t)TGIe)8PTtt#h&Xt?-%Drf;!b*!8[|wA1L yw0dK 0J ]a=?O3b5?%7˭^?a@`<||&x{b?par7^]U}w}?օw;kW$j߃x8"Hf[; ưՓ`0 `!4Xڊv2E'u{3< å=;Ri;b/bCjiN=t)B>|5]?ƪUvug'ijpC<*Y{oR?'# cFu^,COvknRF^PRΫvgܛӗ2):-xc?ὧCmjK*qI!oO*}6]ijoWg}g)Hz]5Ѹ^6nL ܱI~d ϪT^Wê4FOxӫ"<+1-Lv-07xa*űS>ρ[mR18 9 `z|Rxxp_3=D~ˈ}}9 eț)稣TÀ1hZ u7VO,).dcJXej,KW(Inẘ@ IɔkEFs+4]m9煢AR&){l!dDS_{+Թ1<$r:C'%Pwњ& {W(L*_l[;:LH϶(k/YwPJ]U7pu_MC ;<`*ip.vpfU`oAt|z ,6yq (FC[ ߚDR%A~ijHx@<L2fF޵Ƒ"%'i/Ud )$8{}:x˲ג׎8fFRȺmtpH~U_.Qr)eLig}U'1T#{?iq½NaLOR{d;>2,{u1Z8]\ a<7(Z]˯Ox5*z}WMP~fңG*-m[ ].xTA.)|6v[t,W9s`|p`"%sf9,$]a؃׋Q:\yV@`렏w%\>ݠ̨:3Ƙ lho^WnQ"{2.*Y1i+>g:BVD%W#v^r_޿ u_o?* t6 TYjEѻXNkx>GG7 SjO$7^<~W[y7`>*;~+]d,N Yx+ 8{9g/zp iǶwUu.Iv(s{cvc;c_MwQU)_!1kZYNZAڈxu1wn]s_> aU:xO Ԛ%{{ێ*%f߯ʔkv/Xٹo_%*7s*ͮG38n90ʆ1DGخ^$oig:ֱa'Ӝ<:OO)FSuuvϲcxR/wCTohɤwe;K=X 7.ՖyAIWi7qGtv~Tt>{1KzG>fqX$QПLU**6Xac37Ts= G0O{dAaXݼ<{Off Rl}_.l,Fͦ樯QUcD ua6fn-9w N]D!}\[{?>?/{+]Zæ }$B9vFx|8&h\ [* >C2(-}?}-Iә>+?ӨW9w-TRRXl,y] ;gN](N KNe^nTž3{~_AO^B2Ӑ .>\\vNJbN4}C\Y.X|=m4Đ~d8x%jm(1 ]!9^+P@FU LIgs՗@̸=y^Wa];a=ӵrY `aHƁzCI- ޺#s-dmN !ֿΖ ɹWfq:ѥIu?EC0d! F٧ *e-tu֤U.Ib4M6',a}^Q lbR/wlgJ )6 gn٨K/P)7jq&)dLVW%f{=)vd O*{YuC'IEAE mܘK#]A\tZ$JQ]TS{@*C4˜cjX+)ҵ-ŰN{䄽n_LcNXM!&\qh2sO {#G%z_9MOy"?0L.N)y :ZAt4.cۆh0b"A(-U*[Uz1ql1*6~ R 䕮%RmZE(U- ;dmEWJ3U-hEBCʖ$[jtcdը]ܸ(WcdIftUiJ4,4&ʖ#rJl"KSUEQGljf5dHbu.2Ie[j\1gf>B"]'+^U\` J@ C9VRSv!v  e URYWjyth*S2|W -)`J\9JR// TԔ0MBXQr]ƢC IX_N ^^oC6^oC6,z#Ѧ]4߀7RDwT.BAA9h `az˅oZjg芲@2%>AmDʩc84f2GyZ6:o:sMuE愍l W+QIS)SR %ؒJ[2EY8lԚ&ywǰ~dߠf`圃ˠ.'kbQ(4jUp!cqP}+ !z2ZS[C[a 3;|cNg 6cCV>PJ|vEQ}d YʡJ&5yQ 2*۬EGVM2EF=_Fc[U/oj,U(< b|ƆJT ֺ6ҥ+b$B&b`R^h1jx]l\1|k<{'fCόb|g|Tj]8? yGr7LhO^3`/]J~'u mHJ<w',¥E!YRh-: X+;Ap0KBmO^)"ت* oԒx߆_[j*$xǓE$ >:-`.>bJb0|n*RYY)d-ޘ$*O1[ jW"Kv}^cl:m*"mj/W}*]h\v8&sM#/(YXQvL0ER.r(C,bx_&)37CG1~^htʧ 4 s'>g4_1UWsݭF(y97y -n1<#R,PU*ޮ9Y3YoΖtۧd)sBP\OsٴG~6i!ȳiw溏zm{PMٴ[oul㸦u>nJE@ SO. nqy6{h<@\tT7^,V:qz{`\/|ỡg}2Z3!`QZ Z@Ir@Kr?O Hxn~{~HzGKSz"y uQh9t1)/txzL|p|?3:3O gX`tffxUvn[<t:T2yJa_E(@0'ěooκmV늦03x9cwoB(eR=@/t K_A.P&w(L$ɚe~!yDj̆tCS/UzSWB(X`לRf@o ,^ߙ{~KyJbw 83C`y-3F1B{[(@p@M,堪kl6XVhl:޹ĶR\zw\q-`_1L5tsUS~kňHk x@/w~EbWR_9wk $|d '&,üVlφ5GTw+kju(:ÜJj)z+'T &E;iYX(YW@nэzMwGo $"*D- Z5C'd7W JXabUYQfM7SFn +^^WiF URB6E3Z0GuIMP|)aI%d#Zbp앙 I *,3DJ?LYʕ@gr?|k#s@C8-'WUZ8}*Ȭ6Fh EnG Rq֋mU!&"@RkG_Ţ*oFTf4A0[+_K[JDx>^^`z/Ll'~R=9:Yk{H(G"4oX0֒RL^%3S^>}Tx sXm{}pcj7-[jBuD ]i !v (N̓-[)oZ;-wRo׋9%3'3Y#^ h\CiČSѸB51D6Bб6u5ŮS8 .5G)FmE:8fXC*An6fJB1]lE_cʘu-QnۼB5 !x{MQr=߀|"=>xqurt͓.B0% m3uM8 v&(}Sxܞ8L!aIQ90YRrNu(3Pז8tO|0 pWA|E7w,hʆE#X;z=qcu8(9 cМVf: bgn94y=āN9UAؑʬ9dy]ɇV>O9U'{tPyI&sٟg2:zUY(3F<`E;/[2gBdg}m|O̖"/bW\O0h/^E->EѥCå6/}(ÁI록mG!nCDwD53T^J< VAkݫ@Ijzz 7lEhv`Rr"Q'6)6/F6&4";;1O9&r=$K8 a:m3; :g߅,]Zw1 U 9-5付| LOnjr@t2cmzyӎɠ8YFR<h'˱xvs ]Ooثrٔ:gvLؔJi\/YҲSt.ˏl?vńNNtJZ1hv"`5O&t"-_7u4ί ԎA:1Q>.*,~Q|Q;d}{>iV7>PgزPS M[6/^*S#&L.*C)κ`BtbB9B=9~~8hsqo?ooKarP罉/>+}㋣2+`ߦlr9.VJHbJyGGe5'kX3+U뚝z_8Y353$R峮>z}Ѻ٥;r>^ArӍd1Nk~ʼK=88 qwpogGW#ȹ^ BsHJ0|ne'5%ƿ^C=`Rz={ośIWEV\KxuT:-2jIN-h'zs\"{T.[O;~.z-.b˙b<r,FP?rC?) yT_^}1xG4c99DH{;?kO"K*!Ъh^t/mIƠ$h7UA9cjɁ"ж"7o"n^і;!-Քyp6<<>QodP8lS2kL.gNY;5^CBJ*]:ܔ6X Cn+Qְg4,&. r9{/pJ!`h#RCP8ʤ/,7܆ #c\.\B4c7f,"YY`aYo8C1kM=XH%pv͘ҏ/OOy T;S2Cnۼ;S.Xѳ!ء@gCDn7fˊٍuE{-zM(Do4VuISA䜒bW3o%J_.qUP k"C H1B`&ZJd#~1lvkh/d?Fem_U [ ^,*;>Nl} OVqz#.VybS0_d>X>%+bAWv(|?(vTPՆe^!]:893cB]tއ7~Z>x6z dK9ؼi{q/J>_^YpIiW)9,{u  3J9d3dPh0fT3) 6}g9 T!&iK0Rs Cz C'5'oTdx Co}~%uWh)'UAQZ'mYldo &pre,/u+1qd)/,31YLo, ͋ AdIf紽wz'XcaÞCc?WpfLI}1߯VS)F+VAEK:V歶9)Բ瘴Zu㪣$ V7jZY=ןb>9َ(г=6nPh&E狪zmaӶ"5 DЙŧayj&rd[ xJތ" 7(dgK!/ګ[Ծ|]n2։f@*Q\!R7MlqW-@AP1Ut̀6mJ9J9*,u ?'5ebA`s~+goN|UƯjPgzU QpU}2EV;WaQi JMJlL!J/d8j1C '어ɜ.C$s>ikc]я"`FʇVu1!TLXPjJ"}4,rjڋH3A{Ol~ ڙ7eeҗyb+x$EH]荕ԅ0qgdлkϮ]^-;Ea䢒틠"b+bo٘¥̦bXvZ,;Hl{!tR5Ld~W%ԧzi"t] (b V-ycmE$F@UBHlaD3يP Rͭ{(&mІ~б6Y?̦JtR:K(duN`ET3MmUCr,0l37'( RR׍r= 2]:72ڼ_H$2q0nLDl,_@Bf QxR87f^8f6/担σ`W{ZPrhA%Ԙ X)> [ŇVuDM]c*F+P^W *(/lՍբeԲmZ VS*ؘAoml}+0W(ox6ѺI?C΢3V̽ Rn4aocx6sК5F(W|c:؎F0 -ĉCSG_OfB(P,Mhs*&*#n+`/jd@GdJͧDhBP[ؼ[&+ (?ڌڠ=sh^z;W,Xj*ٻ8W %Hu#Y\`vds,*+dF2ܣѵmgZ=]d}$E.!_70Ov=f 2.t§KVY jpJ8uR!QhP2eVXs19L`s,4 5V4ƪvn*-oi84װ¾8~&cWTY{P׼5=+~Kda #G(lЄ? y}͋B͜4&7\sOr1jg^-`.~lN{4t8U !rΑvs~sٕrc}v݃=ہ*'G#AOZOdž?;FHzXg_7o;H-wilrxġ*%ǢjjWh  kug&*Zݻ px.(iYKQ >Zu{)Jg|Tt5a*R]V̯'LqݘV Na)Lb cpLҔ`|#wdZЈ}%'?[Y^buk:/v^-Ȩ{W𣆅MC} >;g=׿'k1gv:AxSLxIkr N HmQZ07 kC&[!L 'F'M9>揇;A3^չ}O7?9HQ[=Y}XQ`xM / eU/p)]ý`,Ob LJ\8r G|ho9zO<cJS&nX{|Ev>ccC?AƲ]q˃34|_sb"{n*(ggJYЭG+8u*(zEEÝY?nkg> +&3:9?9KǝTG *j}H~{qW!hcAm8(їR,xӈs5LqZcFM wB)(D.f}9L(xW87&&RMb 1 bBՃ5En)`,f2,7Բ Ŕ*WƢ]SoDޡqoN$^Ohkdܢ@R!S,w+&h16C0XƯNgd*'ß M<9vrcTV]qztqpx%Qc02YהWW"jX06wh BcN{ipk.iܝ'&5L2HI ݎ ;aW Xs|tݜӣ7!a)I}myv4='&g*U f{y~l^l-C%,eȤ[g]g2eBΐ ڰTGogYg.c#/N]N_ހo^ 4g_dR)vy+D0*j!:~?f|~9 / fv 턟Kf//UGG~~+5zO9?<9};/_''r}g˭2 o񷡷=)DWg:<A@;df)v"שP9:S˿ĕeRfx4wJ)wn:C/HD^yfCtzdiCq%`bOC\muLl Umv(FF[|깴D%Y4WbbXǔl,ɋoEhDhsDPBNJF2F>0#N'Zrmƺ7ss߅1nZ!x]5/fuY=umH^ޯ%w1QP*~)4SmQh&|i "He#0b'-!Bsvy|S'߾/H6rɠ|rIrI6zBY"T#RfsNZ.ZV%G p/]|(s)-% VdqM~bm3CD$².DAc)Qik(UslfI D r.*e桦R[8:W8\X]Ze"b"6KMg" /Vl^L &.g dTH.!Nwո6V_<rҲ|)ΗˍRЌEt,(ZF"\cM&^X:d,z-l¼}/uBYu#GI-R꿈, 3&m_ /f?ݝG _)6d|^]bȡLyфuIl%r\%7FhՉcpѬo8Z?6\ (c)PCɊJDիUk s11oA,oh_)PM\45fkm.07!J!V = 5UnnFqBYSh{r-5G,t4y {^$wiZ' 9[f6# TYgR>+M +E9vP!/ۈWL>&;Sf{6&hoqeQ%BkтOloB5ZAL f¥xNuwoFxas ?fw0k^odaHn "2Oz}Dok^Sp՗/KV *+Ѭ"rA *.ZþիG5Ge%s?Q,ڵ)H#dBpan.Y/V7c i )[9M@Р`Pf򕛖!Abv(&eA_Kq lbZ Yzg+òZ"pd[4PА7RKLB1, td*R*qEQ~@[Qu;QדwFF5kJA`D2uw22Y>Ià8t}8߷ImJ__b5ٮ2ؖ܈q#&{mWp ٟXP;9y1{D>]7 V@zf]Kr@acf.vOK=HYwv*xz2NI~gVIAh+&>~99%rDT&$^UjHP H4((d{Oz C݄F#%Vr 9# w' /n>|2}\]b)7a2I{W<# `}uTfwM@.N~Vl38 U]l,V&G0hSYdZRkcpF+#ה-XCL]Ěs9٥)YYyqKKb *ZD54-mz)뢢x9Sp%AG&buB9|$QdĞdBęsubu:,#$gښۺ_Qqv! 4TT>$F8 )KEph)ea?y܍DM0 pY 5_6J7 P'Y*B.ʂ 0(TW5hi(CC`ԸX+7*ʹpBZ=1['7k^ 8!h9c듃[J< hNKG*͖h]Ds__rW oVx*m:!H~8\*M dMb2?jtuBR9zJy8mki6r/͚,N[ӓ; Z~}-%{JJXwl/8`6b[{hXP3c :xmʑ{͚OyGg_%`nzغ▒}dl-𧾵c|̮T|H>ib;7r$ "F%UAx AA4F.TփmLhC0i1J )eËiL_13߬BkxzZk!nּeg ;GQڱnmn(zc+]9 fjϟ~wFW5\6y-6śv_ڎ]{hF!xhr7RAtH $j<4y=&7v!~e;x^8\װ]|g qAFzc~oCɞg6M՞}[T3F%.ϯlրqViLhDZWFX:ՍE2 ǜ3-b"Vg5Ũ &{4fBzmAHBߙI\f#tz C'\/SvheТ7euUUOTmCL+;mmA{h~[ߖ3B*Ib`]ɅvyomO= vR?dslJԿKyAz S0#oIkpv35"YoּiPԗn/mo)Lh82m{ h`#}BvRZ=jJ>ɴTz#̯wk%P"^IZy9M m (MK%隒=tf +\ɬʬFO_j}[BN&Mr]Gr{.ϧGEe+H{ԱZ G o=`Cj>OlƠ6✕/Z#1x"1k5w}ךUVCm3Ykp 2 ׾i5A[ i ڀ3yHk5|h1{T-;hDvL{(*9ٍȮϾ!Ӛ6Zu6w}?;^9λ=KΙA\܆e}sW~t:d>yQ|嗟+dq:7^/[lgYz3 Ä=c9͛ӎ1LĄ}TSxO3|9/>8c/4?o3Dol|Y (4] ӿvWsCٝv^IPMI*՞b1ߐā1Fܞ U0X,Oa:a;ΪDk<OLeUQC9L4 .^Lؾ<(m0]\plT^}CT:KIHz3[w@Z[Ȫ"=g^_zf C>o \ӫ'g/إ%Hm=$EAATtl^*)4uwLXJ(mҖ\?tT$C:rؖ!qe*DTJ{nfDwSpR:# ˼I$L$l6q1kCS}ər5Z{Tڥg"Z 2JVR\t,llL]QE_E0iԭ1yH'LVv:19&,3ENW?̮t~?k_tz좘O3vZ63?ɇ%Q"D{Ej*z՘ ToE$+[-fV%֮\ctCDM$=iښ߻d2&U ű6" $貭lBFkra~K'*Hd%j2iEE"e,[Ʌ+9k02 c q=dԄ 1i2V +shM,}LAd䐩siVpfxܴ&?qCě2>dwі[ >BxFCJK:u+`LBXXLgV!\|]IڈvɝIC9HHmDQ5DF!3J k:5/B5ePYdfH6_L><݊dӹ_ G<ؒ3͗ךVS)J~5\xkXs7!Ʋ)ךw$Ľ6˱le|2ryJ镫hI BW![?˩?L"V-/ cM5Ix9e]P'IB9F!@fL(@6_GsˆwVN/mctL1*R 2J[2C!ϩw$> E>oOOSM5w&"0ag-'EZٍv"w&V52vFݗaႧXE;P{3TK <Ĩ 8e=< W7=X{a}ugӳ ,,5, r,P?DIA߈j/g/4Qm:k^Ȉ2 jfDP+ݍEgF{¢T+ta.W?<&'..NLKګb&`o),0LThOp2@USb&#:ƞ,Q=5YƘlH ioAЄmuW-f Fc^5X('r!T-.{74yjHK2eWƎ& uI]rb35U@1br58U|yKDuėKw:g^/Zzr~hu(w˔Iy|ttPc(7$L򧳊3FP7?|eWV:rbbbvoCDe ˠVI؄D p։ro,؝]8ב]<7fزGE׀8h=;Z|F~>s`| IF'm=ho;=$ \M' nּЍwԆ螫fEo7 n)L ou"Ư0^rDLcuE*J@[ %W Rye4q8о[fE5v]Rrld547bئڎ͍ƚ><}5}~嵽a?ΊM?O+e%ˣq^t\wk7`^,)5;#Zm%`5/kit}@:mn)Lt UXTƴ3O1mǥt'mfcӇBf>0i 3fJ *m`/M1ztV=.=vxqN#㈑g15q *nov hsvd_q#+a)a`>,ٛA& ~&^KIɦ^VS$eyLǶ,Uu_]^)AasOϠ%2T@*1,뭖}“ Ʀ:bQ?n~i1~ 8Y`8CzpϤGuo DuĢ{sZđ;_>_/P!4"ѡvBN[l8l/!eδ g ̝N%'Ieit82ȨىWvO>URֹU9۔_ wn)MaKc;\h NCX{k?!*! )xkԚjDI  6BZuuU)O>rc$e2o)SvL$}_'!rBuոț'sw1Kz:$aҔaL-uCxrݾd2 þZ1ę-*SŴFS"e9SKiJh*;|WSLї5l;"SR5|hѳ)-3UF"CҦZI[`N#V4T)`J"S:_a.CH=B`ϩJ?w%Z|)v3FI.($`BS 7<(%󌽁r%Ԉ R6HGRCj-v1=,9ۆI[U%8swx mM`BT]qN&~& `˽kW.h6@pZC(.{x#.{xnj¿_@y}}mR(C@b OVn4m=Q5 疳*[Д uC,[けeC!\?w6,?J5zWfh1my:T,0'`@@%Vzip^c ZZ# yQD%1[NN*iIAQ'+@ ;>.6,9l*Qra#)1'Nt@^x#a^f tZoT6h@FB#BaI@k"wF)C6Z,@ڂ{+ŋpLW6 n!yĴO]")hUIUARnӹVEQU2ϜK Xd<]4T{0tl W:F4 yn^Xj$ZXDX"-$E1S^iZC٥x$#LP 3&h 2a8^Kj5Wz*,r"6Ϲ7 3Ts`I[(Y^ h01G 2:b.% }3LTP$>h"l 6I'u&E [0N\r`t8M DR1@@!kqT͈< (bm끘Q?aX!%|?^|Mw'eawAif 6∡]Hi}V&xY0 |(K"uY H_X2p/gI+%PݟqsUͦ܏~%\օߌ>OJ+N´UYi.noxL/JFIj'sr'hڠڇ;QX1M9F6'~Daq0_p"9(Mc`1D8PVО ^Kg~TpҶ>ͻ/l24X-ULI/dh=K:2),i}O߼{} :ش`Ў>6OO&6HCsvRNO5t7Wc/hR묠)NÔWz Sx E.V8ʐ(n+ɧ3;IILé:3XJA1ljf)Z-D/w~8%L`! yI֭lZlP @_[І`*rC M6(Fai`0Gak!X*Ai`(5~kQU8dU>X )"9ucK @lC< iPyYEH3:TSx^>ixP.iTH2?K1/Hʸ|r=>B|mEM Q04U?d{[lbIb'ƨV~#&WX2%7a1q v.c_hnGה6 _R]E]+9nz^XظН)( ~qbpXl]-?m\cq>n=AZ1U͚?PQThץoӏ7hTn)i6VeVJGsw7vሱan@-TPn%*:’a-= C%*OnFyVAY1bw3,OsfE̲|񡥦OTɷɲW՞ϼuMW`|@0Rv+R{\yMFȹ'ejH_g ErYv'~5=r7߭\SyX\*p/vKmTM#0q-#[|h$Q61d9J #*0Aq#)|·N$h9k_ZIs6?oc{V>jk^w{Ѵ(Ü ޒ_.㏴vM|»=ތ7:FK$DGnXt 61keU^n~fwє5JIH:u%hWaw|r4#dqTڡSC0-G'f,/ÿ EoޜV8DXpC :RSq %lH8;~zDízkz*٢z}@yGs{1Wl{Uz˷5ůWqsƗa}էۅn>=8_)fޛ<ۊoee[MV?Z 8LJ[ʹ\쿗>9%w)>5Ei툉(zKNN?ENiӵTku ,5Ugnfee}ǖ`<0zEBH<8Js!ؓj3veO=5ַ[cյm@w{Uˠb8~>| OӷEH;_ EYu>)z>[Z_'jC/!;_<f3D=9:](g;Hr4QkLc ur(}an55g>U!iOoĬ8#CQ1\ASQOpFS?QьtMV S H#!lL+4Ǧ q1c`wQ]ar+LoZ0QG3M}^Ueйt{ H8Nnp5zoglqI߄ ZqmZo6KG  ׄ*F(1<^q߄f^tz ֨KY =sשϤе/-D*%b0IONE)6Z:JxY\Hb"iA)AV_j/Ug>C?f#;zrJh.3q$3Gp^wbRӔx$UWI_%I|UoC,HB&4L!h ׀Q$&1wR" ACvw*n9UpfawLP VAF| Pvv aLPN!inط9%mЖMC2Q$$p~ fD99T:7ofBH0ps+wV=fuvVkuvO儮TJü3 L*4L; 7L"*{Or>F7?UJs)Vywٿ썡#H_H+`%匊/PTHm[4IJGl쎟dY/,T!(Hb K|dP;0 LΦ>1qɶ:b4E:%1JS´n"tP),.[ , Ҽ a;,B6!$y:SqD(Da]ZN^M,60sxjN+n ]XqVYٲ.>e8!c$JiiG4~` J/Dzh"mDϳq,y3ڵ'-M-UqV+#T2:9h;jdLqJ{E9hDx55BbuBYֳۘ\L<M`ܫiv Z+E Ǟُ WRzO>|tL)Xÿ4+ xYʞd)l^f^Nw=nImvC8ೝEp 1'5,iX#I=("$DR_UWWuUWQa{r=X/ C'cǁr"NWĻFՆYLUp ,k˛*or5cWÂRƓ 4Cb= ¤>; i ã g` LMITZ5Q mm闶\O٠.+{˴ U39f%ʑ{!YuFّ˶n^<:Z{Y&P ^Y:3Yբ7O#9Paym}\ĸl%B{wONDևe-hմG8Pz##EN;ѣy_\?{g_ýw翏&Ody%^G f&&Xvh[|5|OBy6ݟ=}6ۺd`y[č~-~HD|%lFr4/ǃO >?!z۸;gߊ߾>0ߝ?x;I#?{|~1棺>{?XL@T f2j,~gz3ax0}ZԙAO9wes2g~.#MwJOb`nt=~_1?\&kE,}~7g>0K(g3ĸף磫a&\o|kMf9fעɂ*^%Vld}?7;%H?ˢS5IH0ub8 i ӿg0ur{!'X@$/KQF8/s3}|Nk!K½=˶ޞmp= X1Ug%.vJӭ.MHc Nd5FIL>$)X{0tbY7M4IO4Ɍw ae4"ih7z"C,_0_i$OҴ3  d'EQS=mV.Z(}0LQ+)J(6a&Ę imRc2El-4Y+$ɫ0K>{Mi4u7WtwӋol@%;Tyd^ԅɬb,.E 0IJ8+Jn{JS2:urߋD\DT{#R4TvK\ ':vJÃ&d;ʧ]n& k;&AZ ^w9w&#\WWhGg y4%wKU]/k*MF@ ~A4?B3+&x'|ϯDi6"Ņ6IkYcCdQF/)dfFg_VւV{HOߓyDO>'-Ч <wш3P/GT'0 Ne̳tU5}Џ8:t=c9# 64#S*ɞ"$i@jX'f|򻌖e.wY{RebʹBԨZRynQDAZ :R|O5* C1! w,v.eun]"w"9(\*d`mIN2Txઃ/2w4}Q#h:wD? RX:T`NHR. =ҋe>XHp%^Q""7KFRg2$FȱXI{ @40>+d'׳ -Hc4C9pƴP?d\$)K$,aDSXe 78zݿ!iGg`?b`'`R'ٜ"/s#7"k Տ;׏'8`juf:yh--Qf1՘F\hy8lP d9`  JgQ9eMj">Z eJq4-vc*P9GH6U~G(g7Jr4uPŃ+`@F"E :ˀKls sW0LBKJ9!7XEHQzy.c=O1\dS bp-d-SDrXE je][bݦ-&뉊'@rX;0AiXw^>~JO$IAE $ĝwPW;%a!%&J ;E5ļT҆ V"|HHr":3pkiKj`HX"pgk~(HeR:KC!SĶ]M%OtX+ɖRX+ ," V9?¶MbO`|63*RwTCAaU{e PݷNq**j{rܨ:|꣖ZH,.еnmYqSJ`w"I.Di0e` 8ءvsw }> @`Ҟ=WFF~YW [Aȋ-˃n)_ՁYg׽*b^hy?yhEϽivAiZNiU`b:5 Dx)̍̌GT'zf_^7)t`85\>AkzIv~f-*kopK>?=(Nu^͐&O^yf31o±{MZb7'0]%™4``eWɗ" *Xg ( s@<Ϭ볋:@&h*6~t)Fp[:fLxj}ώntW52I^ ~eL"f"FpXf1Tbql kɃfdER ª^!3X1J5?.u[߿5۟,ٟ)lSE3UqZ&mO $;DfB/I1;AzTNPqY\PcLEwjݍa9d]FLwt1\aC| b KʼnAkX1$;j %i ㊜#QJOQE Uhf<z.r]:]4"§$ ;/!L 4P)ӄj 9lTJ5\X 厰Zr[p@LkdveSpo3r`rs 9sKĈJ7CIu1cxjwu6 ~Ck [] nr@^?CXm:JFPuW'Hz`|N/cq 7{)}P$¼2S9lP%N`0M<(D˜d'RQQCe-*mVopKTF+aQDQ|<|4d0r[f $B2+0ץ4x3ݓM`V#z#ZKW묡}ӰܟtŋYo}5.'Npst+V{/_̲#켓Ͼ n4ry~5(^IS]O N9;pQ Ν9#+wЌa9,6SKvԍe;M9_3_;Ġ]=KQ`8FQ,V1nEP?VBP~t7Dt1j5:I:X6ŗ};ZͫX.ju*BJT(C 01X`60 fVSVɒZI&Ïi=pP$.c.}si^ߦF8X5:DjB:`ll %0$Q)SSAju3&G>)"'w?/Mqt&йEv$Y/k;'E$@_f|D.kF-8$w %WB$KFs߯8Z͞m?53=Ūd=c wƨ@I 9L9. =B{u#YJ1[Ol9,j?[ f}P[Vj,Ԓ30ܠШU;|4]>5%7{Gw.(OTU2Han7@+vJ`lm:]q=QpH|f~Wa>[ 盺;/')Ɵ4Jꩧy!woG//OM{_f.0Z|]g$ +N{t|&go% Α#t+kHpTqV 8w/y}˻ 2g{̫ Inъ~uj+bzH>!b# *h.ءS&s kW?CRI1hAӏ0HKK9ܠ oH{?ȵqJ]Vrﲒ{ܻA yC譍$Vr_X#$;urV6)8$/S8eߊG%˧f3Ʒ#\vVxO0!q=5ix~sVhfsɷ;l.؜<ݎ}!H]E1* `2|庹{qέ>9T% |>l>}a\È'-ct*/U~/RZȄӅKwKﯞ\)D$:KtuZt7sHT4ӹ!֎>튺_|? ̙dQvP&I2鯁:{˓^SyFv\`F‘-X*7;Եɷh2C,!CwȐQ/+, #hZ兑8p>"䛨9wV:ic >%iޱhZnؚaT+ג!P3Q9󉷾ehN1{]PUPs!{KkOj<>?^a-3'PYy ^r_YW=jIF5CIdJouB27dejni%KwM#Z2? & 7Q)6Km9niLl5zN >FpFZphIDXt{vMr! R !R㤓:vJHdtCK!hhr ǽ!W36YP΃24aޓʢٴSrX.NY#'r CJ ŒJ?̨۝˷zQ1xǙ1 3Yק.OzؼGrm/6{?m !-JL4$'n2Ɗ )(gFS^L|I##ЕCq3 6HBiJs_~,@E-u1jV_;/u3/>IJH&,}2x,!N;QIUဳN^i)Jޚ|\ YCچ̽H1efKi-1[Ekq>x] g1n(,[ߢVMJI" {"=jnր?sFA8s3ad6la4F#y[R'4yXk$/T1$.7TB0Em FZ9 9U3s8q#ؗXƀnoi~2IFi Vjc5MsAj[d%5Bے,x,IvnɥHcۖa@M(t ¸t[TW'9[BL4g}E @7Tĺg̬@vh쐑FIɺp 0&M>jKo[5bf!NC֎-R) rTK@s{J^5UdyV] QU*24 z$250\ +|9ʘDw0 rT }-%T\)MSél5@ &>J?nRmA=R}q~nیQjy2'y t.ktդ޼Ra_-m"pSn<؜asg<6yo-rGnGA2Œ(K/_kWdVD.ҋߛ\}(>-:\ !_9Dk0% A:s߳L;n-nc +h 7Gokn[-JS>m5LUTBLA̷>߭]nzNe8}\Fy“ӳGVܟ?,̋2]3x< |<g#,s/CQpi<:*jQ eiFެPj7IZ  Q-x]]߇`<:n乥Sh0i\$PiC1\ZӦ-.*&QV(bȸVQ9q)bϤ-& g@l:S`Z}r/r_,mG=lGrfٜk=g:R,*2RY:rdXqO$HiھoUiCɕkOY+uoϯ>,ׯθЧ~׆utރdO^zm~ 3͢Xʤj|SdM)kN}N4U#wAߊug0;>h6dQ!rR8SHr9 &*rjZ`ة {r_ۅRCbk7H_StzsEz0.72~Mў2ě&%Hj$ңayByv_7Kics+ػI}N ;kz? FBW5H*\;gɷlh&o L Q\4>ޱdUdS\ntK6B+-k d<0.1L6&5M~߶HqBR4>"6aCl}hl`qNk GAkƖVGeRK7$XH#]ρtxņԬ+PYKV$ E<9F16薝Pݥޯ0=lЂ"H <ͣzt I4hÉdöQ&f/;y]Ւ1|?JޞxoPvL`{fUIBJb!Qc xҧ\dщJ^0L>m(&KϾ)@8L(aHkPٸ3Eu; !Ύ\6e=u|vPٔC!_yNјKZ)( # ʓy5 2Qb\gѼ|X)2q2W$ZߠmI! f8rjE+H,Գ뙎܏NWϮ 8" ?6v'^j z= z&%/UYu(ʪO`TĻ`T02`qOţ;y>Zi;r ȪU=c/5;Y} No/ΉݐK/z^ˋwy./Vu+77WM]oGW}YrT!N6_.ikRT"_5IY#3Ñ%CH,szuuU T;"NV2K9 p-CI,r8s&?gj/qD<ɅHeF ׾h* P%3Y6OC[/lx#zzv$,0<%[%J {ljr61%ɦ1%k2_k#Q[[f(w3/73BD5<ފe7P ,jR`v%Z!_ŭO `"E^5[W2FC=^# {#TE`oUcywp&_;O1Ed;1ƭ~q'zm@6m}۶e)zx-r)vk#&[TpPO;p~~%p[0M/v\zㅉ\Eye&`p[{@_C"ܗ˂F0Ŗzrbr@V/\Cm~;9t.{b8yG" BwmJR>2sR 'ĨTOJZ@P{It;n1ny5n9BΠSڛ.E?ve?:啕q~eeKg =Z` '[M$2(K$^YEo Kg`:5hp^n{X.1(#mڕ(کk_~]͗xĩ[ڨ:M^JkB8Hb`v<pt;n1FM5F BwYS1L{jd+x9tD(Z Ξ^ѬSx8b뭽G}wNIe׋WacLlO%<]'|zŸy{\Hp\{L~L+);[du[xu cr`8Z3"#9,}}c%D%gDoNtR4'8@62# H~J{;OY>=< 5n1<ݢp@`0sos5>*q&Dm26.(R0 kUsvðQm#3sWw X3ѕDفJ4{%Zt9)ݰ/-ƊN &pD% { MwDDig5+Z-?Tfs;kиAofEGv^/w:ڇG{GfY0V0)@ ,PϨ$%k-$D\6: āJXHdH#*ר;Yl݅K 2\VCLO.p\_e6z4yw+6rVv${zv3˛7oRPPx|1ڳG@&ԞMjE 20\璛 L/]OU0;FXИW3/wwҳZa+{mzjbśS UAP*A Yy xg}0&p팧-\=kV% !+ƃz;6g\|z;A`NXkhyocƨ*xrKgJ@kdt6@Q~ImB2 xP4O>yDr@׎Z4͍L2eځ  :j<\&C IcLUC WՕ\Q#QY:iKKK(8%Yý:W |c˚Yݸ֊,p:9x?ޚYGuaӓp<Ƣ]Dǰ)5o׷4XZկp~gl?ѭ@9E?EV'fp}OR5E!q{z6F ;k$ZGGiŬmI8@JD!f &* [-=2ĩDYDXsR4Cb+f}oYW@k?lk]!ABå!$k:0t= N8$/U3k͙hu5<n"11ܑ1\ 7V\65UǺBެ %ZeSLt|2Cg4'c8̮UW>Q@cx?s k˛6huw䳸-2pP3AGysG3O<52@޾ο-=N/s>Xއ8oryz yKbM)H&Yn62)*1(Z:I:z8_zB:K=Nӧk3<ï|7_^tGփnBJ(djv \hRȦ&[j({Rb5Q0[J^\\_-bіw|%0 Z(b~-OEFkz#s7#o8=ekXe<ȣ%"w9Z6Wz t(^fnU9&%,AD卵 Lņ8"M;:&'/:SՋr/:>aP{R3W9\J7{ OVJj줗]Ys:N1AzA2=LNY܈DGb ()nVjpãB/] `q9VlK~Uܣj@! k-qzp6. &[mNUP%y΅q$WrXkT M[ NV3Z$+\H !Lr Q k:BZ.b7PQ =DnT>XI%=ѬZF'n@`ԅ0Ci$p2͆WzwRzx*4׆aw|v->TڞU=,EI3[}`#ц脻;E"cOXD)sv3AbK4N_ƩFTF$>1܃kedQY" 7OmuK؏ P}GicF K e5zPvNC{9EP?w5HbLg]N ?\rs8Lৌ8qnyǃT554̬xC=Elz,Si2Hy5qXĚ O)B*>1Is8_Ś}'Z'ٺ޺ }篔6B\/x5~<=8J\x7L~ |őnN/>lN9¹oOΏCfpa^Ͽ_ɩ/ gGkpsC&+\\P!!1(skemЮ $h%= 𩂅Ӓb~;MWiDkXCBtdR/hȅDs0(Y}PG lZB`Հ\}[cJ?+ ΊZ;TIzP$‹ oAU$#s%%)ln&^} iy6bרs zNe޵md"+sitI/" m(o^#}PD")2Öhss9> m uAHep )'9$|#ip5Gk$-3H )ŘFI'dDy6*.tW!; rc'ggNRhO$MOR%^`v#bX Z.R'0ߎR Fq])tau;bWХ< Ji!rqhFG'4]x*w!#4LOF?-m)"' YĴHGP1va s 11c.[B1& NVcI<4 =HbXL}9k! yDĜH}\a]8#"2ךkԇf`H eLVB1wnbXv\~x:`;s9FNTQ TLсy)7W(|۞VpO azBa2a;zMc{_YDp<9 !(R7؛!\閨呀·:`J(&R(*GaDrkOx9:NIiy(?('n6Ns oVKB$d $#lX"ai#2;!dr_g$~7ב3AíbeQ:oQ0<' \*9^f(3`bS ynn"5I)zHuV(%DnR&0J`HT?6DT#i4 Y"l֫ C4EYĠ N 9*یNes!YZ /%hkp6iZ(bo- aUJd=}<;9|{O(*dc`EasF.X yFSĊ{#lvDF Gc6<m#yg Rt@]xaER"D{>[ݖD%t-w^Eh">+GܻaA\枟e1G;"] w-+Fg2_Nٱs`в"hљE 9@㚕6̣ Z݂C#0WwX6.BJڼˎZT[޵rєF/a9ΐhBMyֽrJ 'Q8%ǚ~M1zr#-ny$t Cc>\U ͇kc⅟w `wX̒uh1˜ 0 @0F? T23'y>9jqC3,C>NK/-5v~z'rײ՟g?? 3x8>VSE2Ql:CC D?rdam2}{lfmcwa7OqAY7;xoVvRۡ0dm2yzGIKډw}[?XY C20H_FFBώ-eb@T8K$-Hfe,fJ"Jb,G ʹN2A$Q3odkl BJx=Y V[<F6͓ź`JpU`]CE&JRn2f]{>)ݷ~>Dup5m^{gyr! KoU0c6ަ!wXE,әmggu{3aA;nח;[Q[h 6Ki9s Ok.qc.Iw/^C$ sJ":/^cM>)D~|Kbf-;i{O08 B{ym q4 K.bL L&w9E,evʟls?Rb  &,PDcT=!F""a!3I*٭yrM`gl@XQA-v*#4^heVLD(cyu<*NI\Ҧ_>sVxQCO_v@ r)bE1z(1c~r"l93D&G^[( !QL9$HsK$ܱ,ϧi0 !==*~ӎլ.":iF~ZokPIyUBPV"DD_@`-&$O. ]hI1íi<֭iђaZGS'" 6 E@hb 9n冲喓Jm$hj]]_.jy@uNsls$Dkr 9M_n6t{3afy2>vUӟ>߹'YC'a&xe*DI)j2ƬGΧ강f?病j)AL.= ƿZN_ IBͰ t$?pq{~r ƀ+.EP />iijŐg)<(?)1Qz,MKɉٶmJ^wJbM$ZhdIr)Azp #ˣs:0?.$hW=:z"]J}ْBҎ3I\e,DLuUТ%?rPAǸu|;>@B":&;>A16qu" Ti;ߏm/a@pyR#^䬲3bS}E1}'[G,(;̂8`*Iu_KenI1#.s2?Z,8yW G?xh`˯xlQ(0r^Ig }scɧ?ÚL;S Q#A SC1PQYyL4^L'3z xw.01Fn"Ot?'N(uE>֦)R M^f4f2P޶nwWޮvL $VV hnM8%mmCA$Z Dyx`Xfp -K8[p%Ϩ w)0j"1X]IJkiAҭama݇a#x[oHl/69;9Lg5'G (;Ԛ\.5juU5Y,s9ky9mUd1ap2"D*ERf1ьubt2P]}2%3,a/N lx1stm*|bὝޏrxW7>~-̃`6cfL6դ>c=7wS E}JTy;&f}?I|J;W厇„~\JOm\yr$ "2mVtG&MnEyp}FvALw$Qr.f zu+CBp-$SR4cZ4uS7"v+ʃSD3X bzSE.f zu+CBp-$S5c"uc֭(Nc.$2~ح0Z2$ B2%Q8w[C݊)<֭ykP .f zu+CBp-$S4~-dǺH[Q":EǺչiIb֭0Z2$ -'ʔ#Ex]fcN"4%AU1x ag Պm7ӤlN\ErWz>y<ו>VSEYbj% p/W:?pﱙǟH{`獏On8qȻ'zݯo^ 0vw߬\?^~;#:M҇'~~ots~fqs8 G7Bi$/##pgG kGrgS,)9-Hfe /3,Vq!@p{$-> j56}raJx6* b4O? eDp^s#6Z&Juߦۼ ֋?Lc3ꦴ{>)ݷ~>Dupud OLi^`OC&fCbE5LJu֮ -ڜ\b3rE-%jK5!W$osF犀x(vyq&^CBQ]5'R3񍉮FDWIFW͎46jNt&mtUrti`'D7{ί'K:W=F<9A{ ?m#Iapm!e$$ fVQ3߯II!Eh-"z7ݞkfGFw\? uY-`ZW2s7K%ӏvݽtϒV 6bN]ϩo]Tb @!֔Wy0-BhQti*c5Q NU'`d\5mxK'nokΠCP 8YVN[N^}nfiKq잻#` ͮ6X y-`b̮Wux;>(_xfk<fn#?xn¨CsPB1g z(@ @=z*o9@K/L|p7ëQFxD˥/{a8xfv^&-?G֊OYkCrc4 fǃ)dU``’? t㨿=#_C9T_BpSxOnCljۋK2|=v~f,& ƥw6؛h0=t alqBvxqk7gl?ވW#g;(k7PCA# 5+yEM>s|VnAYm&J4!CZV26dcS Qݎ]Í:ػ(mUTѳA7(]*_1(0Iؼ4ߑ焞 Ak~,jV3}yYGW |aPOc(Ac?O/Q8f1,xL-)VR"c1-~lƗ & QzB$#'L \r/7&Ƌk{ ^& N h6|oO!/?Zk5haiNf|-&|6_M$,,B\yϏҷ?]*~담8[.\jx+:洧։{oS_ohiVH>DJd{PhhJs-o }WzZr6PM4(8 (HRᵚj3E(g 8\yӠ fYw&uJ채(~iݡ}^:N[>%şv)s[kĈɕ"T2XXMV,'#an"(+$U\YJA bL\`>F9)_1ٺٻkC.> 1M}䉻anY1fn9uwkG6Dk#˖v\VyЀ_/o/to8_Uija oebX_y޼x߼zYmPdGH2q2+)e32kvٜ30էלMC>#ONIW%U>p\<0q~!,(+ަnPIDj/G?hji:Vƕ܅>| @il,Qsc(F Ldդj4hr‹ߌ l58m &fϸYɩt%D].#]RA+kiV)i)h+xm'W9{ HڲV,*_a5)$.7 c%PwMNꇸq*rMG}[jN?)9HlOm}׿]s.Eq,䌭&!vnP阣B_P̹5y+O_Ũ꽢 tt~眖ӳVM]>'&{w3=*#=6۲> TPwzK{y˼]A41}f=]&xNЩ9ڲvΉƂmpvwdgHeIs򃢥ڣA)%t8z?tA"# pxPBX=#!TcE0vSeN`=MsPdBjBFN:)۰jIj,6@J 94c wWޣi!Zⱒí%K+0%3i:6$T@ycP'$ p-Y0b [K%& < Pݵ8jBϽXSD4a%+qa#C-lRSCP$Ʉ2d-\ X>oR1R y Ҥp$(b!%,1 2(%VŴ|Zn2(hPW#8xbvm$..W+b7 2x}[2&<}Hң(m8(ܵGQq17y7gނ}5nogٹ%>j^tZo[c?UU|*7$e[# {x2 gS`[+|O֯ym 鏵vE AEEKOStʢZi7=5eJLq v7/nK/W%\k)maQuoRY!(R̭@q8 *~|F@A0qž!'hKz%8EqфYcjah"#Ԥ<֢V×]3EE1߭5_Z)8>wkMT !A-; b#'f}~`pa03 Ob:Mtxr,[)d{JDkiD8wE-٫{MNKvBQK%@@TF62w=o< x#!:Kߡ^bӡVj߫k2B`#"e~U7>Ү3q8K,jBL-9{vo*q-C]hMtu8Ԃl,stqz.d +ORīm\Dx9gnIԉJj@~H !CѠ4Ƥ#ko˞R0_%+Zײç(+,Dygf71 ^ ?לqIrX'5!4&\EC`_C%NQ#&Nl$&ՍMz4 6_<k& LUefr2mf¯]P6( D;F|aN{j?)sQ'-lۥIhA. £?KDt5˿\ʧ{gDXq9t~lNS$\ \ί=Zm̩3ϯVaGm%(;Ԩ:ԗ2rTJm#Dwld]r `Km^u'Nd,e"6B@捲MӼv)m^q 88Fk j B h!kw6c}4[nkG.SCJ%glCQh[.-YՁ"C*6ҋf 􀲢⾶ENYޫI}8jZW[8b5cŲX^ gRx3= #-dS_5^V'(^~!:2G !$@91Sbu-O{"HK^n*po(fDZ&호l ysk\VOn/dHW^)kJ9|. ϴN(>O?;4i-`wCi‡8Y,8Q\"7'Aѭט{N9>W$3!ٽ-J%0fطuI;=fH$'~$#89San\][G*A(٫. J]tиYMH;3gP^ gy CÀI`6 'j}oEL]8" nk&}LR|l/T|n" SPxԉuDns: ġ[v9IMÂ535ߊ4iiڵVنיԈiRlqˌU />k(s^b`FnzhkgiΟbH{i)(L-Mjpuj^ØuY4[3LE_~< Ҭ8ܫ<σtda ooW&0 @[t{zk?1^&Uٽ@g] cH_ }$|!G9x\w]d:3 N&K.0Іph3 ;!zB nmޖzR8nbnVSdS6yϷNޱk\ #,ʓ>}$ߓ~}UZ&M77<Dt@x0G=&ltO)=gF濅zM۱A PB8 4}LKc:i%, O7>}v87ǭ&>bJƬ}pY>dxtROggB+%ŴgXCm`DTH0 @HE cIDBζaJ lc 43W2 w4˃L0XC{|⹫>"e11Iܟ>fN/#7}ĽJ=2Q3'v$cgeGRĹ*́Dӏ%PMɻfj<9!!88TX ׆T@s؟$+Ĵ[  vF9Jl泱zFٴ0.}dJƖt8 AO9$@N> Cʰ1fA?AVq`Ene!]D97B1&ni!q{'1C|*Qx:STMs3_>4 iSaxF7e>a^|_ˉo ,)ˉ^@ZyoX3ϗd ?R OI5q8=e >R+y-Wl?bUz&f֋'a=߉igq- 5yf;+/V[qW[zĭfS=7,ݗXmo{lJ}Ŋe=OT­j]maĐq:;%Fx.oJʏ eAcc;0Gl( UwᲃpNғ=] S6feob椼$e쒩oTtxI7 H[%C~HpoLz{%޼8]# "lN6*zӬ3j0zsqӌX.Qgq$쒺(B/KҷIa@OSy HM6^FK kw̛."(I{6R$Ґ5]5˓JZr7/"IMfeŃܰՌl8ˊ0fa[6#JB~AK y I<#I<@ zkyɑh4<֧r8n @bJ۹6Z$+YI~ E%Kh> @)ف&F3珷M5>cԜ,OAN{,B0᫇vW*]ޤ[J |~4|U&r0ua=u# 4u%#`t4N `q=ߧzy% l0L&@U8~欇&[ ?Lco,yx} wc/ܗ,K{@+􃐾\A+M C-G`TZb*K Yf腀`Y=t&}!Sxj~DI+i* MaaasJ -xes'湓uf‰a֙/&|'WJԺѫ ʼj#|EoXtG"M"j<yم^,%*)j_~zQߴ;J9]oכR9ǧOYu[b;7e.' 8/CBG4ĀsH09 ( (0"6#~1jbOIci~ q$5~S8*kJQg2J!ֶ3D5ϸ3Bow G)::spk i:eA$RxB!29-"Y^9L  'ZFa_f'B!}t[h}}džaR՝;4[Ӌm!L%dK@xR܈&$ 51%ā^סB>GޱۖцoYP!v"gYR36Bc ZCK1aLr$=ת}M':؂ѕxPfI񴁽46B8 \U2_$ ^hI6Xٞ|{Va}:11, z.S` sml|yӞZDMjM$ $`6s +7vyc>;=S,I!!@׾=֫{s8ZW ~ py)Ӗ?Xy_EKd+ԧK T |rG[ Eo q@ApĹ;=)+ %A4)wY[#Q9V΂LI?C.3vtXf|7Cesԍ` ] p Iebff)۴Z'2HOݦLϊ}SSv’ z;T<p|NJpR B!@^f1}ڿϿ>Wp)< H蠐*nkj>zoaHb\*QzWdY-s}Tz:KѾ;5W;tz])E#CTHlcSUK~a]ѶrP2qwdFEVzlsf*Űr,K|$O?SdI닽λ,_>bU'/I9='1zTQ5 rs}tz?qod槁]+g9~mC魯VSϜanyj=*ϯkqP7OOEopm{Fr4O&V ;2t̶~OF;.}Oa y{^ׅ+)e/A9x{5Ms謵-3'@k3'nIUXqbZ!ɤPA <=nOt:U3$쏔IIefmd3Of3-;gߓI&fj((#1bc3j%=͑FW% l#M|f*17xMlޡt市%Zs_swz q$T1E2,J 0bN'XؙXO[H,96JeNpkA2GB\fJ'Ÿ9J.\NNZi3fS0Ё7%Zm(BP(u$Oyd6xp?}zd4|E_\`) O!ېؕvDp)R%s%oy}[INk6~8}$&Ms8>AiaX\S6w. krpM`RaM9zLukm8%"\[V):L4m [WLj!-kF)1ӪUhz.<Ďd\VxhF!6)=91{ưՠŲy$l>萯k^hYc@_{OSAۤĬy')o8\ǎw <y$%9h bxI) \SD!z2* jlSU^oU Y+5ރb0YF!(\EZ($ "NgGe7OS|1ؒB8mv`+I%G'V"OG3҇d5MҡnNVnWg+8Mcc'y[nFȌrU jvV43#<|W͞4*b04C٤%=neʰ)"slI #5j9&E$` hMg.%R-ݻnm^ ")}}dNY"5‰J S*3XPRZm #(d䁞Q)װ` !Z`%?4*'kJ5N00Jjр0Q¶_ _#|&k‡<;RɊme*$i'OeOUz9)g-Dr605})E7+k).yɾwC$kȁy`2o:Ĕӝ#7!=\֤XlPS:0,jrR2-lQ|E:.Aخڒ<#XB⴨R 6A `-^W"Ւ;:>[KE#,d&R!XfvB}Z]L%U%%OcAj X*d8Ei\lx=/5b4tw+rԞ突.DzIq61^/[nfhe~ Ƥ\TևA LqFTwm ǘRxc{5ߚon:ሙI?Ib &)EyMjUR \)5EE)RCEǦqB,O(aNDL$#2'9fr[Tb/-[k{W3Fi{A>Qxe?htmiP\g<ƩrM}O549-M<}`Kߎwhjd ױ9cXVAlx+'wA4ntY+,c.r CXXdi=t1AagTgLe{[=Ie/q 11]˜n$^3,U,G۫V]P)q(lSGuqt^ !?\DEd.Q˖kW~eM5֢a/hA虥 4fg~%?D (b'0XFP^ԶpsmK糙ft~nPdE/bwen#,BǠi ;?miOwicl&YU s9 b3Fa:&$ )ggDhC Kqy7J Nt|٤C5FxL\饟{ӇN~.6/ijX]-[2?m)[aRg83T&1fTR!CUg([3hEŊJ6*V•+q2%)%R& .3 e~)!6rH[W pFVps\u?o|Hf2**&W UᜓV]hV5&dBPn㄁?nvcg$1 GHɔVRXXAji J 0;VƲQߗ !l.YQF"?K7w"Ow>x}6U~JϢ\uh<?[ZW0EwϯJ &#_z8"d ѝ"7WOg>8W)N:7{!;?h;oEEm_~xWxV$r5 3!&TBҚaww(N{np> &Oj2HXR%ue[¸Lk &.AiǟgdxK5f2R#.d:zc.y.3K|4yJ7hʪ(P T3X{)Mפ?Ak`JLC0 >f}ÉJ%.qY#E 9:=(")8Q'q!4*v|5%DgJ]*PP-RsS;<ֱxol:vyЙIZN^7+xLuu.}.Nh9tQp$}û3VKd%}uƧ7$(ŷW]1w˲^ bW 0{WUօ_J hF,PDVRe˾xWvP?+u`b,ZXgfʰ3y0DC3cPΕWjz5}!-EFek^VX:/ WuN+ap_z7CnkAPihCVӞC@ʮQmXXI}7ogj*c;˨u:9\o=8xl+D[^=[ y@YPsX[7!A oۈ?Ep {S6Ndȇ^AbךN`w9#Zni:[K0I{0ݿ.kyc߁SejMB! wrEhouIp7 ΑErZ޶tr Zj9E٨4KXr3aMΔT *P1fiLrQ6&lB 0-OnID0^^'hW.+t!RYw?k]\#9yw_55uCw?  ;ޭ7nKXŻԍ0'q"zvs ˉ50^V/ k-C,(s@tnzuh6Ma"yNG7ט.Z]?8ӟ>14uV[<ߔ峉_,|_okhɏy#?pHp?O^=aBw@5/g1EbB Ks(`b4iA(6>yٟd&\AM 8p\/и(|WYR=̋Vx—+\>X+PiQ*BK6f_TG$Id= X-7a>_􆱞dP?]#XI>f(kM%<-gZhLMa48`7ol A9%!Cfݐ ҇?{Wȍ C/3ޙp]3c3bBV )x(PUԖ#ݔ@ߗH$Di3jd1PGYNJ9mΥlUbBZ9ӭ iӺ'Җ>RA i3|.mcMg w~14z!8-4(ic44`s4 Nk>=B tfaajS'6h_T(yifz9e(ԮSnAYDw>fH_dQodxɾ}SMv>eR͝H#e`&W %Bisbd!)玉G㊗77luXl_6f1([>̽7["(Ow8 6r vܐYs2Z"@F9QFQTcDrtH]ǮaBqA#f-\nc!= Xݻތ{ؒ@lXq*٫R4Okt2J1"9Wvz9W#ZK~ցGӶ b1i/V}g.*IZfQm>`F2odNb֐5Lp u֙<܍FReOt|Rӷ||rFHSJsCՉe F`gjD!xBVd.X_A;Zr!^.{:;)<ȏK:Ζ h2_is= /i.ސs_n̔V E9_,|,+2VJgNO_۝k*w 8[<]{2A:]w>R}'&1b=(/!0/zjM|~ /[r{0+i>-q$boLSN3Yj>WQg-'?}EΜ鷻ȼz]ƔTv'f3dyv7q#!SS;rtPtZ9IrچP?,b<eKSȷTT";Y8,Vd1L9j3"H.;ƓX[6+cNkݱ~߱ppPcdLK)D2WRJgG8ޮکVk .R \pJ\pPlri3)PsV&;ABZ -6M(Dk)ccVSEۊ:H$cJjS"S2(&F:h0bM (${( 0%{peL$%[y?N\ʹ[ܮdsԛ9rbZ<-]/L~n 9m'$Ё)^heM T2™aL+s.X^rhma@aIR I %ó`i*WH;Xdey7>o-xSv"rf0E/\St`ʔĈYeFkPT]l狲խI5ڒ v oxjUZ1J \[5"}.<-0>ęt0I%"9SBPP6#$P?OHnO?op!ݐ.V)putvl}B=,Y lLr%^݅9UNX[\Z'r198RR$K-UaՆZ+J&]"vN~:ަF5Bl6\7~j?#7ۼ!)=*G!_\Q_k]\B+#D\Շ5 s("{թ(Xfg[ Au . Д\E%;%xX9r J/~ 2> ~ { cLk!Xؾ!~> ďh3o!,'lGh9=Ec/i|]u WO>0߂Z(Β_8γ8Y2T2olCF9m8G<`eN0 gHliXjcﲻٻwл#Sث? NFKT,&GIi,[ʜQ찥F7 9p^`_VaǨ-R\;rAܩ2?R%1Ds 9-cJimJׅo['vqTR{ӶD9e%u;ð\= R \PfLkPz5ͪ“!àffʍ`8)-fyQ . SRIY)LלeHA+Q"kto 5rï;η__]4k=\Cfp&ch*b\V^`k5)]#P?PS7շ_Ƈֆ2GCTg.L3hDX h-L7Ѵܩt\њ(pr~\E$\"8} ג8}Z{`MҲ"טƦE3ISNPc*J:͈vيChdnY->>\!`# 8EArn(bz7߾Ƒ ŏKFLZpʩ&,(K̉QWB%Y맞H-'S@lNQmҺ:4b<^뵨|;E>~ eLs* lŊd>׶s^+Я/QCE(b)*e7#k`m2y, mcDodōUH$](L!DKz02:7)Ty; (l%ŧޞJ=~;[,3K~BlWcoWj3<`'e =y0Պ{ H'~fǓl7|>M0Iރƻ!iאW!_|Y,݇"{NUNݽXdP[ܙEӋ`^IJhѕw^`V}&+AI@)^CLtZ?yv7.ً-@ 51%f6{ p%?r'TeUuf(7&܉kej}V}Rx ;/J̫6kYbeqBI|'JkS[Kl<jװsTyl}({ Nab|RkN}Xb^qdbSyA֨wȋ _F[rfMp|tu?;h Ɓk 5tہǣM_ϣu5=)idyLڣktßϷGiJ OcktjuwzRs|7U|kzdne_QEOmwBX5B {ΐ !;g#~* Ow\%yX?;pyCe[gP'4ףYy>f>:f2; yQDZI)Apk[ 7G|>N|ɳt-),2֒Fg9RO%B eI¸J@x Z 6J "?HKO$A&̱IdtLRpNzՇbU3KM']@FZz|%) Xc> ~$)!yz$)47? 97j9dHN)?5M:xTd͹ 6a0!lj )M!iyrƤn-xSm૦a# }uy? b.Y`:i^SO$MVDtd1C~tJS!`'rÃI QPlHC pR@ Aa$Ф`H6-Ͱ~ ӯMs10̎4taR9K `*밼ϓ̄= h^shvVÀ8TMNۇ3,Aʉl ݴWE8îN8#_/菀T;SL?[>,?샙wn~]}rl>J\R] "%a9#haE>Ofrx7r>}ͦ.<,gt7{&d:Y({3_L/Gޕ5qῂK$>T\CeI~\9I J== X,qkKXvq8J<>.\drћt_P.5F 4=4 "fjÓsRA6]1\r:+`Rm^?@jZvW+Gew w߼703`yC9??yWXǂ̘1P&V "lP[c8 $v% x PH W6N BBz1WT0bJ%`^A0$% @h E-(f$vH`m4*hcq˹6"e#@[)TsZ k81U8W ¸ahR8_zGi_h0=7Vcq}e,p(/GĉǸ4L :D琑( ^Mį2pdb/z!R} 1,8ȝYe`~3D EX+--anݩV qur{|[PbE-*$ L;f1yG#h(bo% ͸_NQevwWk%F,*kFO7˩Bk+PAd(2[ qkȮ3xa+ q+|GloY޾v̅Kg; J~2<]Q\EW?AG'v QN5WDVh#`b" (Fe\pdXGf Әmf+w5.r0]ڂ"XA UR)-zE=r2Tx'N5Rx2(;g$AV"{@@$h3elQ3\0`s#C@`*1(kuv\N[Q 2GGIaχR0SPwx"Tdp@ЯS cbKSˣmy'&םb\ d0KvRfl_&k94)<&'E }%"n58,:ߟ/J\^}ǒЋp? Dh")F -e>˾b3 ]wٰ7賌as=KUJPTDl5!\Et㣏7hr1:ƺ. f5[U4K>uӺUкbPGur߱u;\%"ɴ[~ϫukBCf8ֆuXcnQwlcWL2l-#]Ӻ5!\Est vUM&0:떋A}6p$Jw[6u[U4K>n٦uS*aPd 뀂LcW!]| y*S }WuS r[c;6SdJcLFnU4K<7M󣨻bPGur߱u;*ieFS[lFZ&4䙫hN)u6?@떋A}63/AlݲkݚАgY:%ةEeH;HU'f ֊YMv4pF"tvFsQ:ĩvQ#g݉Ӂq;SHXɓ[ NvVC;ё *B^jxot$h%On (m69`D1hYk;klq=x)}z5i;kmgё:W{ڳ*yYk;kMA =Κ~ZYkb$T;ΚDH`gMRZYkt$]Cv֤V&GB^gMQvGSO!a0 u AsN0,!E/6 ?4( -6 BI.41`:-`ģ2dϐbDPxq&~brG!l_ L'<ҨJN􈀪2:(mtȄd!44CbEdWj)(MaɄP~ZH`sYSiDu*BmxXʤ#с2B%NBl1Ab1J)Enr05\;1! !C++ 0-üREB(-lzU=ㇳw ;`fVu0X~8k2T1 h (F B;J EdˆKH^̍pf8"kB( ITQNCR!Y "`¼cils6AGEY }hw hY7]Uzs U:IyKJ={SPa>]3~eQ:gvYb8RcΖ1 n?E^|ޭ0/K9, ½5Ǩwu޽}مt6/Vb4 ^ЋW;#3/:w&+~Yp"!ڒ.)T5ow)/ I!pW?~-τYi}Q5GVGN>r©" CgO&l/Ӥ%}!p\T]=7eϛp=. v9^W36` ƅ{w`RR!Tg FdQLYPtwbb0D_*QqJS~hFM;w(kVۿB\ne1MbNETk}]wCZi κ7l(#}N488:ØעgD]=E]^D)Td%{+b ŨX9A2},XMY>HiI2q}4SMṗH0ȗx(bE X.! KU~'Ʉ7l(|t2eSQ"b~ћt_t g$=qOx@ /sշO L\ӽмuB( =iNz`DhmpCv1c" U&R99YnÒٿG,)%0G9 h5 \r9d+)ɥgs^if`Z8+v&Wf[?_"0Oɼ&#ʷE$-WKe\IoYG}w!/LAz~8 )45=bqTRIpooa eI牨:/O=Q[-m;hz8Ρx$ .%Iw=gjFQ}VwK8b #D EkCҜjV)>*9ݫ$1Le/+gI PEr劣I,ʡHD̵鵞hK%׿yP\F!#rT])B&|p[_CZjQ{]G;& ሑAD |>P4`'Ռͭ 2)Gc4x9 @@mD4(ػ Zo60·E6 -~`F^=>:#D36p,B1h -A`S%Ǹ6 6OI2!6sE!z+b eBUqŀ` '3#{+ gdb~_DME?1_{V6BQC1kYR!UZHpwDeҔ)digm+H`\ٲ1l&Ҳb2,yWS(ըw;ˁOr Q*J$biL!>w0 fv>8~jϋ"̖/|3 n)VtKDMx|ŷw^#vY~Ռpg$çRQ=|E0-*JCCl%.M 8US<*y:E+rEi|}tkS\y(!yIY?qd z6i㕷퐖5lRIޢvp1s.wYUVfkθiB]%-ȸ0]%*d!#L7Hc|2 V?oLHRg%->uMfl.0]0G{+n`Rh7uhxDu^f76f>^m?"? }xL<Ճ6ݤUœ٧*__rz_.if|H.k^nG3}lڊ`YM&Eϙe/[Ӻz]]d:vNCv؀<8r*5zRRI&&_{Ĕy!|6B1'eGqj">K=8rII0jO FZ9,'P gwkkX*M14CpXZ֧nY~bhIy,6;J+M4N["vHD9t4"B75'bd %p S j3,FɈt"(>r TkxP<@ 2mCrR$"MK7[u7+M6&{ g:}d_>I LDK,Êro%RJGLn>NC V[lw&%=SPN 9f=Q:C/y9QfzÔoCٻFncWXzI%piT'JΉ{ dSLRkoRi )ix[b83׉W|_wh49xu-DmǸƦ}z|35Esn p&B>/eBuL΅|Qi=#}m=)w~B@$d]ږхAc $~:J2+:%kmtl-%h8euy,9n[|]57H6/G JUӢ>']a1ZnzufBe2Z%B9rx7h !CP΀( P(4EX>hf1UyW5r^Vn3G ?NirgQ7EcOl,p< 5N  `)L]!-XhWc+uYlu&!qq6zx_7Ѩ-xúU(읢hZ? bfZyF׮aaOpV]E0S_>g4ꮦ}'crVd>@81'BN\<ZOL|ؚйeiC [@YX :БSˊzëX1,z[ N;a|'&J/l{uy 1KF{eYƖw|S%vJ#v6;)oC_M#Ц}f)~Xx[ѐ }:bugo(>&M8G/w;6s6% +.e&4}&k5)H d% >'U$J.K@Χ#$G&87[(ǧ}a}au ,I*(, JB3Ub\ɴP IV&?;G"Rh>T,\/'quqX*?hb@]=Aކf;X{鯷mhP}jΆko=X"wmKKN~h!'4&y qZ虇ӒnBtG{R& s=8MEU7g~]\)Nh@ Ȟ8E +wi=# Y@>489VӣHn)d)R[s)Аdԗ5$e&\jj *,e))Np2(UN`!<}m*t䤇YwcVVV!Xk4v^NJҢw@eow6޹w2/VE2ZmM^UsB5 Ūz¥^]:ЙϥvK=#. iXQ*duѠq5R:XSw ? ^cQY@ ڝtQ#cI]Q[pr2*c1E)yG`%||, F`.1%Ǹ2 z+.[ }+B,c-X %{aBQYl*AA|mLׄ!_6_5%×?}16#Я $.DJw(%J3AW z.x`H29Xah3H?#IBCIrP>Yn@iP`tq-4rwcP25\TAZs(z[f7=.)W̨ `&TBP9휟7ˠ&Fp*8 `iIa J`pd30(T b?p$6'h7iAFѸ_<-x!}:&uFI PFPB`Gp,׆ /M҃*vl7Ϡ0[f@VT@ :5-VPzLrG4bVuM\@'=W# ͥ5(B+H1vC0Sͪg.P3l{ղDS=GB_^xhvm&N%`A~'H7/NJCfh6wο{FJh\YjϬ*0pP$o@0fQx=$=Y)E~'8å,ŝ7Epl3I\`}xʓr[RM՟'J{;zq8::?fsƀbn~<[;`g =]8M*[n1Fk>]Y3,/%IWVE IkrQS7J^08JV`xa WPk5:L\-/:]>3g \eGlQ!Ő תtZzaU4?"@a0.J` (Ty=cQ-`T 3YbR cΔiFD Nމ?ww1iLuq0Z2@z;Dd z+Όswk$ɿVj>\c 櫱O+ A{Z8{s%y C⿰P. =PKqW Buα-o~ v)j( @t==4BgdIFz9U U3E# 2XEV!咆ܑPa'h;=+PQp !zASe -%R:?CM ps; {BS~gc>O[sVya|c$'FpQ%a\zT?!my2H4{> P)K#tEѸT]cZȜAGz:7('إ3e]XZnӛxMuw╈\+X@ɠ0M@zD(ew< l vaNo63|znAx0TI׫O>KkcjM]&jN:YcL)>0]wcQi;{6kvɊAF齜9ۯ׿$7ʨb ]IIIIue eJݔG/3B)fFitT)ZU^fhV>ھ|o7wuwӝ/XѬ -^1`sRRf+9 8j"|sځj̿_f G|yr&Ůo%I8[sv[1X^V7=V/ר Ck[1)Ohy-= !0ƃ#LXCwJG'KeBmtc=W!ar%[cB͵Jz5k]}xrٲyeE`D\/N&Wg?OVOdz%c~~8 -_x+߼۞U}Ǽi@0#TnvLܳaH?_)]E0WN,isٺ\zC$qVr^># LLJc+ĎVgIOvET>F*ıX򞽢RSY7ΕVr('P .h-{F=i{3ԠX+sm@UBj"^#N$>WP잽4krZ]zx[@F.@3>2w̝ps7gTz_C찕˃U '&znQF×}JELC ɼ[Hk {VR>ҟ/K<垶k58o}IVZs8kߡ\14Żn1HkŢS!ǡhWB0N~u聾8@~WId:QyhZpY]6w/Hw>oQ;*ޕ$"igdGfdрvb3,Qv[Hr7%UHQat,2+;ƕ+ޔ^7vlH_{=q#kolV[m{3$i}u[ Hj 0kEc,eF0͙C g8NW&3o:|s f<D]nnb p~Yzsuu7~{bLv' kT?aQ{#39GY?9MR[MiyzsNbTK A8|}i1h7yyR: f'"쎝s14/NpMGEl+nxw]½m?N5wwꨀ#Rt֬<8#ΡA9񸥍[G gȍnL*@˫KBlƸm2)wY}~J=fd@ѫtM_ӋXiho5aʉ< %+7T|{{\^б@њS7:si/q5%3.ʉRMsiB$4^Yi6/G2$_V[1}Xhԫr_m1OK2/^5Th>`y]|_VYj8ƍBЯRvԨda&m䐲RbhZHM25j0k}sX<֦ߦKlB $#_AB/ " ZI^8f@['hDuTv+ll}Oj;VZjΚVm+Eє@AJ:笋/Y,CPrfD}_ԫ}ߗ/n#ı ҽh2ڸ6&/&Cv 1N ogO ]vrK6:+ieX"7DfF/hɉ֘\ Z8=i=˳=}hk&dI dWUDYAh!%X7#VDI{KJ,j[)v ZyY^/NAÌ}WCZi:eBϟ:'rR+q;'> SzbP){қۻbiv_\f8pzro6'`:gpY[= Jszr_^^PuZcJf5avy85%vϭǜcfL"^hN%3_JCFgzH}t923љU9d]q,vC, U94B(N"J:J 8B]+*Kr([V!, ww7k;F|Tݙa =Q-SgmmfN~G iq'ʌnUE{hƣ^P.IIꨊ s_Sw'|l1P+db%cb̘Y+^fXҩqSh&uao@P YvF:/NyK/ʤ;YYQgĨ1$aL6 _S>;"_|s+f: 'Wy 3j*>2zI,(2d;-jcٸ[7/. ]ڽ➌7íuBjuCT6'‰0QLbEcu+'O.sjw5f3cQݡ~ϯ|74bw,n>ہA`0Wnw#Ǒ{>zmígvn1Q4PWBݗwoo#ΠƲv&[RT|9}:[_8vDT4]nmE "O\4QEt/T͘@Ket:!1X6 &!<z&&?~<liyM4xfV411RvԆGx%Fs"ВwF^)ZE>q(P|qiԃ]݂qKǼ%}٦m<(9njp,QΑ0fxsG>qI] agt1ҕ;_φV"{9FFt]: #3hH5?R=ҐAj ?CqW[.z|dͥ]>?z/ʒG!-~?? ݋ŒeeGHoI]Z3=\Bq|WK#c4^>Kv ,(ɿ{wx65uf1]ҙYԗa[1I hnZ8A"E٦Mb/v+RE7lyv;;ȳ Ϫ< }V! Oxeoxsp2%B@fy"@vE}t薂BHA%(a2ҩA}kO"mnY6ڴ`LDg\J (;S&f7oBvJΝ ߬o5f1Xmu kD}@quL'YeVz'c)tpgԖlWP@D6Y+ۢ ur'"q]0#$<@̨" S C)cTΒ;b tY5Gz6nO) GK^gAqF Inm|X ? XKNJپ fBM5cv,$0riV͓ e͟,*q1x*gr?IA˂<>ȦPIP֦,%NzQ6*Y8wA4_~y)7U6F <K40#1:TTLD)itu8V5=l/ùlwII9heNkYT[R냁/ MGwuLk/ذ̩5KR(DV#ZW$g}Jos֝jP`u8GT"4ك̰]7pX,-׎.:82Տ~[@@fٗ/Y 2@ZrjŶٌj\NH1;H$r3/ ݲhY@!]D!r{w,;Aqd 6D\D9`I( SbN+=#] P@D7t"2%3<#%]"}HsnXI"J 2;[ 0ٙ cU(E fP (v ϩKEq"0h3dGFN019vIdyX8Tf)4?di ˠ-BKԅ=)g,d$%ȑmYE2$eN,_!GZlHdW׷:b!Y^I&#'6G]Tr6=2"әmYlPplg\ $M$bK(C(2{Z3Eº-](夭q~+)X$g6TԌZJ~Vb̐ZKub]Ѹz 'j]\c79@i n11DnY\K9sS0usRԈ{~ҫk~~M>_mrgQ6`m1 I}<LJy)ږ^ unULBg}Rn.X4-uğp$iI٧>Y!sqCJm^'%61$SYpxt ;TAqZ_^|NoPPԄ3hz%8] ݇^慢{s1 IE^ }1wx4pIMH?]S@HQ;^jgX6/hmڞ&5py.yo:xi\zO!jk+apų+ :T^QIn/[zE뻂71Дߐ;SZ"% I^q/zIT*CP?1B $$ỳ|#X\US0#B(YFKVNVUPlia/ X~=r14Ɂ:Ei,ց=nDD~|nW&rtqL khѴch8ǣfu[xnz5m-އ'KS3UZۊu 2YKg7QYUm&k"Pz@N{K9IK,8!9έZP.csp(\0LMwuA/Lm(bkbD6v^Rvkaͫ2v5w7ê66押G< F']7{HHLtGX8(H iп~HKǜ( ,6_wx 0@!<6D8Ib"ܽ$jIT]ܪ *:8ͷ{Aݢz@C j͜^gtx$LjM;O᳠%(̽WM%*_(QMf7\Gk(*U&fiDދc\/u&L&PLv?Y->G*[F5sK-3BL"fWXS+"i;+|8/fَ(`JyF0WЂ6|rc:9=1Up]U:Hb#] y StP%L#Fg{C .[SVP.Az)$͞B}7د"}W}boec ԃ1 EP28w)B"#r%696X-%KCъC .gH%U)X+K0`2kKX+cA9Xȹދj9LND:6ۊ= UͭQҰaQknTId|TX !HJ R*?c*<1C$ U^ıC TBV0.\aTK]L 6CTIqi!*ϻlA1{~B`_Q* NxvdWmBaSYwq+eVyrNž/sz~!z8?DQBx[N|:Z'Y%m˲0g~}ӆm{]%mPƝنe+eωf&g*(ړ,8^/4*So^/[O+ !DSEpn;5\p!mT^GɃIz $ifu}-6`9Gf^MA cU9j_nY[MQƀ> _HmZ,G jќvZIHЂ牗F PKMHK#2bE؅@LpYႻnsHVzCo>vwWjTK,΂b?ֻZyW""ŏn}JԞoqPtwO)$@^v:^ȃh-31O6P_CiHk?Y]K&q(8>v:&\OQJkte3sM 822"rA1ơ kx-p"Rxw57݆ON Oqi#RxϕO!}> k$sx`P#*Yt~`c¨hLnhro,7T8dtnD&W֘\\yxF0ˀfg가oL[nV $`L4;g7K,D3Pk=.Ͱk4Ra&;!m =aR?KJq(]g1HKMQTU dppX)"8t4_1 `1Hj׺Qv;hG.P\<7Ekץ=t x ֜HA!(` n$u@zacDH1˲K1IVIU8R(|f$tw`ř :ߐ y"2B\\2f3r;Pr/76)QBWD*i 6;Bb%\ZfjB/ ƜToKxj*_A%=]AՔ¸4ѵw;K2{]EID1aWپʲR.+!?EwջsU>nP7G`TXҔ9;,1 Qk<$ ]4?m!M]ZkנD!Qꏪƫf#%VkqGw޿/JdCh۝RYJ\y x:yyܖ<ٺﮋ~p] p9;ϧ{wtn]!4q)t[v69Fܭ MD:h{(fs3󼞆HFZ&ĵ+ƒ@P ]j[c/@X *:7k/3( 6v$HVʞm8IkNH~ 9[˛6ݻ׷ɷG`@قiim#9\шɠs=+\E%3j"b S:׹eRri$lXȫ6W^/V~Kwas g21NpS F[i|qol\F`ɅMYvЋFƀ(QyDUӫt!k*{i%#>ӂ  OvSVp@ă肀9\QeFLj3Fd{p$Qk*v8-H㘐~lݼ}܎&%׷5̸y9`}h4&Yp͸!xYXGO.p &5- V=+akAWb,5>\8I)?ҷ5qJ3)"BP摤~T[ Ft!,8 lmQH8V:yPrN3eiE! zƮ껏%J}U tsusq`$RZF $ݔUo1jk¡xY-ܳE`Bwolij( K8 a}(ހO#䯃D)a37t!+cge8bV'R^yÁ[?-8eN!#(C@>͢9]i24dSD5czRB|ms46 O4=B_Iqv6X | ~ʷgz'"TVpH/ w#A{쭰BKÕRY¢3sB IpgI{F+B>4ffC^l:vwN{A#0u؊eIdljdӧHQE )U}ԫUtupܗ^ԂG( f7A>qya\;m$!b$!à=%<,i2:’V&-8F{:fu JۤD?;d4BqS8M:H2+`1Alw.b|Q!,%I ,v,r+zCp(.kΏO(+Fyؼ҈V؆YbY4S"6=ɨTF' {i V2uA htʹ*hC5ӱ٤ZՀFHH IN6f6M+#x҄y812c%_"d8ɩh*5իGgx<_D(YN WØEbI!A@@m8Ue&MF\6%v۲emIɚ 9W`Tj8,F$L(ISdu6MH:Z%RtKLbl0CsT1~ PRՐmK*A+mSSsR"^[ Kq y,͢QZ\QuAٌk&! f)R $,zv.b(]Q%Es5x0) Zd,K Vs U g a$^|7[eAj%/mF?ĊGB^Tm꟤W3-Ѭ7UprzQӊwr&z0\b V3ӠNc*aPœ! ֳ#ζWf6Sx 1:'qe9}iZ-5pȐrǨC*g_k,0exVB()X +heB9rT2Y`D*Esځ5$A{:I;TIʩr06$ X`)(ȕe2q KɝD#&L]+B6ag4fYN;N_.l ~$=i?E$uF`R)J$5z; jI0:903 I9rD )lE@?N8f4{9nkw :k͎]^@lae T|4`'E1;#_SozڃLJs)PRoO;(ANĻqsYj3/b<:`Brt֝GSReM.U_>~7l'G\&l81þX\z~:i Hs Edܛ}@ n%aYl7L{C {,+δ S:p042딱֖T7;-`%=%bP<KbX* Svqz+"ЃzX@ے*[ɀaKNցC"p7K5ՎO/jA!%Xkbk*IXgU`܎R)y Qߌ:K8(w5Y8H%C856tQ_p̻7;w^l~fg Y{v{7/~}s~?޾`s@^yS_LOYoIozdziGg%GЯw׼VgމuO/00w~ slg G?4 s .eN2X ߸̾}peUQqoo?5ӣf3<H֮s-r,d8&u|%y5)d[᭟]׹j{϶AVs0KÙ=w`~ o@~3FȨ7 X[7._I{}Oūtχm0qlCf78x;cwg# ٲ/,iy23Q~`gbMƵ5~Fρl{PmOLX;<Bo/Fcu{_@s~; i'_if(i݃22RqgZ,ή/xz6l&/7Tpet'gz<v:L0HοpyrpmcR$J%hmHB3+vF{1$ crm :f ߆x]؁q9$+ ]is+(|HI<܇AV+Qv~bkN E0$$qgACXĵ=<==I CTḬ6*Al$DV>殝XpsU ,8q- Ks <8l ? 8 g T0Qmվ6|5i?WY|\eH[c=?W뿝~!hG1u|坽%)U:eH58 )%J*Rsb,W_X c)D WW(Lr"{we y4(ѠGy4a+hq`pF @%4q)/',d&$#Sg&RK|8 ,p!.@Fs݀f`m׀H} bӓ¯zjt\9q')"MhF ;eIbrB~Jcz`fJ2uhRJ]xEH0+TPD*PB **tE'4}`\jm#T>9v黸h Gn:eY!gdՁMâ 3Ov-5Ʋ!^%^Znc$D&4cՂ-LW FiW._J_}_TX;`/??~ՒP?CL/X?VTlv0h}0?\<ع\]ohىð{+[{ RW'E741'K5ʹ9Lw`%jU5>^S_*}oݝIg \=[b?PCkQ6Hc^3A uA1pm|5qrW2u4 Tn11a#G!dUo8]jYE/ʅ|Q.|[vMR3v]’(M(D’M۷c*BKٮ lHsFYY]|8yF":'ǂ&>&Fs<6m0a6LbD+ pTyԚ"K1P-24br Udr18 B̩40&Gq4Vx }Vu dUԸm)cK,pW+lKKC715'"PXMb  6<%b g3ށ|)8|[nxMf/wpVq H\;;$eQAn h4S3XW{|ʨP4?[f~YXwM-4K^D f`B+dZV6YA&cQ`qހ%6`Md =\G_DCJzu󳟽{7\OgW{Ƥ}N|o4z?]H9$q0kO ))8SFzUy ˺&˷z(P-zKMZ2A y5{r8_vo{3ohw?߾ոuxxp5*_8 [<i$*kL_2'og4ގs<[C>t9J N5bל>߯F$xb_Eo!{dqVŨfeBw ]aO34F|>nCZl "go4lH5ƞQ }像l}8Q};sqDbN.jۇj&5R}-ePx͡bJ.oXIjf sXZeoy8s R2]I+NzG&0dRp?] R R|r͞m2[V);-ɾx RI]6!*ϗj.R@90ٻ< wҿ\'ft6z gxᵟ]|j$jj0w/.6 Cp&_rJ>Q腰d﨤$eT@U ʙ2>2E uϢDU*GUaɊnr_ȡF-QSkԷ5jiZZ֨Z;uSښ[=5j9f~΃Νʩ.ӛURx8cyjI[c12CJ;kG."WҰ\<8 \(1K':M?<'s\) )gCI?qa:\F%@K,?zxUK }-G 05NM`B49JFMB+5tpYQ)PvBmƯ {Q@ѹJ=*ЇAFgeU漬\\LFƅ\:;Gx$Cx9\rdKZ)P{rQQqLٓ9=S7`Dm%;8[)PRQǜ_o)P3z~嗵$e(LhGߨbWWdyj4Kog}jMKp6jd S}[ ϗDj#m[C. q(h5L9L 5lwC}9_9A3PM]OA/WڸWQVc´ӮУYZju_pIstF&x2a "'TJQE"2XDy Jqg_=[tv;Pp怾}^^fy0|?ISgZ8N$j.|P.h%VߪJX ~~p&1m : ڹ L}YNU{-M\Ԫz_>'Ҁ[H=67zHQ#dPa0qSl \y %F[-Xhj30Pȱ"b-fDBʤ}&FSU#ܿ 2^ j;38a*$BII5A(2yM-)SdLޒ[2yK&2y7?/G?;3yy`Q"NGFZI*˜p|_k@ۛ*h& k>Z`X sbEm%,oi4 XJ(m%g~{zSrj%$=}b;Z PruBg~n?Lc$J,  VX`DS^9-bb4ro  hN!>'RK?$* -Bj (;p>hDJN^KrL9J"(f!gG `=2!A.@M]^au:m?5.' 41:#p;*IFM# ǣӉR@\SayqGps$h0TXsbA]\‚ *,‚ǂ.0촓AodAZEH7wu=rFzY`;bɢ?,66i)RE{{4QmM{q˶i5xn9EJ ;* ػf?sx*ze" gW *x37=~t} oߋt&9&`gO &zFsv8WX} Ϲc?6me4[j sj-V+ݙ,غSjc[.p9.~↟GfHw@ܭ?<-k q}0a7~ytxa[Hv) xsӓqӓqӓqӓw?}@O&:zj=k]#Ejl(Jn57H15oڣiu[Gbd^땡?3{ڎ>?|5G;;ߴ>"㲂W&W{B*8?,Ļ9?3t78=v}d|MwΖ 4]$ 哋,a)_H2{F=ꈯ{{8ԪL %3ģ,ge/+dZ͜RK55Sy%5"+but2;OonJҟ:MOR$y.T~ts5WU7WYK`˷gPRl`s;|?}T>< L#~w]{8X&@?{\*Rxyz8=3W@&!_W,G _PìFdpp(y|-?{Ό@/3Le9Lʕgc Obm0|qL 1B~r2K]lL 0iǛuj6!W%OQ0$73wDلIe}[Ϻ#JS''$ž"2jha9pm!FA rӖ"~~;ȦZ"7Sȿn I!'M04Z@ȈwW d|mޤDp`1ڊvImp& ɿ(ذ<"OG9 :$ Z(l-@>&lQ iag\8޺vnLD7RG^i쵟^i쵟?1[woAL$}r~<1Vez@zV <STLu<6)B;Py-qGKVxl2HsJ&Oa&5:rn3t{قr6O#2VNQbC k9F;1 sl΁/41fj:ܺvnL|137^%y +.EܾKe_H)D}6CF0Oѱ3h$R?Jk[.RM~VӮ~} ⤪: ]e$}4)ˡN.,F˳U@7P6[ZraS QwG̊q Kqzh:޵jdԹN)N\`ovgRSAB3Ild0}܂G@C%!4]*ۘMɦ5^o Lmo˝OՐ{y}q M iiwU-t֦ LdqrtY$fvZQAfp̹AfR-!Eu`9Q?>MX뵎YJ>B Tk)):lur4WEH@&9K u$4mR!3Mk!C_VhYCTPDiAPCw8덲X a{%!;%1F["OAyQguL^̴o|;7}3H |#+B9D"j<9 VACd$R"D:r;YP0IWn};7&ALN\@ڮ/, "Džױ{osWdAųW8VBADW&sφ@{ct?xDRL޸vnH(tcn@ﭽsE!Jh]$ <!J2T{s~b掀7%Qһ*2S0zog dtxd$c*=rT2<a?N0 DD$M 9N3p>@O%vnLDPa LT7+˸scOrƬ.\>sC+ dnlpCz;F$s`1΁Rá3A"C?*I>QdCn\ 69 6H!}v'vjز*R_}`l,UěUhw2}d7g=g>c&)zi!47X]F:~Ù0 W;;tU*Mv+zO\`vثEPH)RlGhB6nnL-Rk‹;se|3xNsǧӤr͡J ۀz%;Q1݇ٙt#um9s/lj}86h.\8xO},sђ1V]MVjc=&?ѼBjWZ 9l썥s0Uy3xpf>Czo:P^jf`KH¢c'lTY?vǹC ?ثIߊ$9_ 37!á5v,5Ar{KHGv%? t ~̘J.h3 9ޕ%-W*r2xrݝJ} s+Bf^u:b Ʊ~GṇM_z k+6ٯC|<}_RWB1u1+2ЧTlr`tY٘r Z#d@IR` 0|z%$"fEaJ$yK閭N`i=-3 1n) AƳ8 * ;9ƛx?Yn};7fhǨ3QI{׻tpEw/Nmm2LWe/ ?%ɶXYmt"ͱ"~yKN?w:-_\#;g)( 4ɎB{mT,rTꇶe>W-78!XjEmx͂02d+jZ7&]TBE|ȪNr NH5 R(mT洔`-4[af$ӬI9'5ј*C0E뚤z%~#zxc屬x3Pb\R.ָjJ M38֜_bAc&LZ[rOX%`-C-;D Eє]F0p1[ Mc3܏M3 e`C>)fa`?&z*.Lu ~ LSo~p*A;[{ O~?s-|nQ~ͦc~/&+n %35v_UWcۙwZM9n]n^.+,.]72=CkH @2N|k&G ()Y3~txDE 57;[A:0(هs(#سv[;'|ۇ?le&/_\Z0X Er7f䗫4~~ߚkz8$ iM(#HRj0,};B"tmi?֒"#jH!f56q<7zTȐqdxRE\P*IKQ&DB4#'/n#cMi rT'S_GGiV[i;XyŝSyFzGEWTG=G?Y/J»^Eà f,RtRoS^ќht&XEXJĉKBxk vzɅ'-!:qr63%$LjOs\FL,1~hF>#k?%Of';e:ɇfU30/aM?h1Ex>i?K)e)c%_ XU) `+| ׇ &tdֿ3I<PI~CA=tmQ=`giB9gFZN3S"Tδ SrQ!"u><ba9LU_UYݕZB|16;/n`~2>JObPcӼc8~\0jQm 1Bז~x\㜉*]*(N2a빢e5s$S4^YlSN W"]O(V7V6؋oVQ-Ur:x4$~|[C $F TL!AFE[c:X[Pe@P=C?BH"Nw>ϿǹO5œ6v ˥\(ɌjR˳eݮ^CE!/J%YcA?`P }AC smBhbtj _ə=/dQsj@`1scyFLx-!V-{TzO vqR.poQ%T=x֪'@y?,u**js H&BK9ҩ8>bF)&V0)%Wid"1%FbzA}&0h"%X6Ħ}%ytE 7/d,|[Šsu} S(^1Z&)MMSXVd ?f ޑ9 (.F_w AO,̎>Mfj?3  pW%xYQS _Qw^D#!Boz\]^C%TXxTK'eɝz|WRPFtI i}E`<6^N|"c5M4a^`B* q u8CvUv)TZ^[dƲcjvm 䒖 /d[zgE|&^ ڽ(^E,\ Mz> ^<+}jrLEUn?KPði0WWcDy.gPbDD7-jeY9j.tH`}(xjC$0zbU]P`q r=GSkFn>zD ˯,[.f=),'m/w6iND>r_*R:%N|vnm_-ޯk:IeiRT:Z+Iqq8TSڬU;Tj/YL&'_-!dZ!NjZ4{?6fL7vx8GѨޯQ4uhE@V4=ܻ'H;bD]Q{a+Z}!USK`"Ќh3NBK" :ּD7?gwU%(%1mL}2 zGRM2YXXPZJG4w&_z nFu팁 N E+ڢt-%BD{&r2;(.{AƬ*kvO+efյ#=<7ݠ`tIQ\*Ty8.vDvJ:^R9r<(uBćK*2# AZɲ`WH@YA fٱ7{4U[IXSe*wY;_c" O9]"Xvrd@3sA"h!Wlϳqst_\}d(&$^Ҩ8޽V^{FcosC\97^G"9^iVQ]Q8/9y饹15RJCB%84*tKJr/Ł@ h^_ Uʎ$U*M mɣeZKky4p"#XKwȪ7ދE_#$͠Q5zk}j%M2Ii Z"&@Ӱ3.Mq1ɤ,ZHDŽS":4STÍ;ߓ=u,ȸƖkc({!œ0;+F0tXYy50)YȨHOqԔYN$c {)3:?zv۰ކ]_Kᙔ  `75*R! S҂cfhZJՎ*R-p_yeU1{H|m+9(%ln@+Ի(#+*,4$q|&y9odSPs.6Q)O+>);5RD l5MNyJ$`Kȶ;d7k3+%'wW1%;) '< <ҽEf۝#|Tw%bx[7okfuݢ@Ӽle:{jaNQ!<]!/qb{"E 枱n{׹{ok;\ͥԭm^H72؟1pp,`1´cky c{~{3~5{}7 1\fB@s y~8Ds<`-9uR Qn[[pRνo!kO:&wP 8AeBF@C"Sq5,:d<&%*dbXFx 6ROa'c3X3yQ"~Vh9RI>&BJ!@3LsG1jͬ)JUpX8d5xHjUXQjTG ))n?߆ʼn6ṈG ]Xn1:r"<[zK ?~_0|Ļ_؃`W Fgӏp6ϟfpqw1Z}{0~gE5aT)%cMb'`VLjkNY!uǥT5[$I˭z{w)i%'|V^{ ,$R'ŢRiP`ԄQ^cĹ+nvEUQ$% UxI1XĪ忒qˊ/%dMĻqϊ)!+Q4VVPMwySc2mM INpbRn=,LFI4Ѱz%k YAts)EfOq(wZ>*CWw";R{\t"f?ImAzGg b*WWTл.J.cikOUGv!..BRʗr[|ob"cn<q\*KEQ##(wbukh0_t;jQ;(9X]CA~gA"=8QױurTjpBD. sT3Z&ipL@`m5_8Uncq wglDch!Yx=$o=֘sbJpxv0j+K'L'P;e%t` |?(TJUqPv1hːfV6U.WNͧ`7> Dj)~f)Z`'f̘,[ke9bت5%6 ўgbm(XlB'kdE$[ ` WOC@T-~yb?9p%MkcYKϐb8 P| =DkK!JG~GH-sOR*BzH\dnU*"gR$qD Mł =O cZrJJʒ,p86 2GoU#ˊ4oAJVTd |^C84LV_=ݘ eSp_ZX"Q0ԁm ,p9Ά:nufĦ%Txg8'y z8# t8|$ٿϏ_,Ih;\?,6MR= # %;}鼾P+QX۽Uo_Ũ u<vI*\r:f {?$;2U5UA`Yv<{)i!-TZ DSƨNrH\Qi]QJeoHm $+sHeqZ,⭈-H]Q!ZG `W0͜Ϳ{+LFCdiJ}WӻV)Bv2~8يh+zjޡ ;6D<0QqDB"yƉk_aE裴I3}.^8c!q@ X'b\&>1GXFvSٽ"q@N.1Jtw[J`̅b{e%ʋam{.1tubxݽ4vwؑ;WN׾7P4Rv?V7V7aD2Eͫs`yC{_$'w}3HW,9Un,yg_?-6F);mlLekЗ.EO.4FeF;D6/`%&<6‡DVp -4Wo6[w%Hٖg%{O8tT)OrLe10xPrNbU~ ZcU6m9?9j<9S4~\8l&Lbʀ酙ܿ,Q1S+ۻ} 06LBR7;-nWeuTeGuBa7-VUs,BEdkx`,B.,vg}{WyXHoTrS6xCU=PXN PD6mN[Jh)Ծ8\v\B8.R eYbm(0^iIOIDv¢l)MfٴP+U_?\.Sp-!=/oU3lcxPίN_=pm9峷Ւ Njubq7oō uCqS.I~k;pmV Lʩ.kY]eV,kЗ.^OtD2g69`cŵKt}a~3C>GTq#+FٿINTSORS c]icDk +4m4ԅ,}CeN B˓ 4q} 9L `tr.v j &lٚG-3=? }{0Y/~a;|ŭ` D QMU7/[窭nZ䬣2̱vLJP2Kcsuzq`xk%”5Lw& ތgb&TRzbd%yʛȭ6 {X!;sqz.lj|7 V:^JĎ-xx0hq ڔ]`ݱkMX:a GAI# gﬔ~+i-#o)~Yޮ'Q JcZ4L\2[2?C~IrcIZ,=, ?kiV̅oKdOه;Jv8!pu ;/::\xPϪV3n7MAOoFh<67%X/nԳ=zr֓.ɺhȜdҹ @j;wu{<䡄_W.KG"1>G񇉗<7}e4r2-{PG\Y])8+t%UMBX%D- ~2kn鬨4 <EWdl%RIh3OSxnSXsGn" M4im)'p%Kҝ2;.?yXn8sN"@Q2p~J>е%nߟ~ΞX ?y_qr k,i~6;<-%7vZ,]vzŔ7vGn!#!?N?a 32qg ڰ6%(ƌzQ?A|#RRf_^"o~פo(䷴{t<wV~>t#&evQqbƆ~t;t~=9 +~?@9A0_pWo뻡'mI,J0. qɐҏv&]ϺY-b%*)h0=O_rQ?|@asvEJF({1vvc}azo#si"Ti)Gl3S0[9ȓa:8%t=}.iTKVO.fiB|! N (?}߶Ǭ[eh`qAy>jRf U/f؛POI[V9lNM1=Q‘쟎uN+NƋ|*AA]aC!@s(]ZX3pn(>}ToKZ317dz^4UXIf[Il`uf`1$[Ԍnߚ#LOҫ'L-}gZ~8}YO82ǧzuzk[CAZ\Wxb68]|:}b-1h=3Ow곯% wZqטa"%Ѓpnro-zƇ1LOѵ}NH>ɄQv-%D-1=յTkk{s%FY- #VzW%dGkS-fSAS Ckֵ" NCc;xpVfIw*|uk}m'ִ`JY}[Xt?,iFΥ\v{AFR7`ԡ kxЗu E[k8څV$Tй D}4BpHQQQq/3oFl O_aJ' Z[UA;r(ʌQ+"3F[UpoR"T=1}]j "\[߫"/G(wr>ܨ)x70)ٜYCR\=IM\wػFW!Hr݋(sn9@l\*kIjTϐ%p8y%rU]]W[z|A?_2&~pA[Dv>L~6_}p-nnC~5j3+3V]^wp-6#QL&L!V,\\3^GQfghڛ*݄%,à^h򤀥7Ռ/paG /N|dP*p"Bq0b_9SOZ+5YZDA(Q>gƫN 0"~nn>ܺͺ@"nbe+)KLRq9lL~ݛsE-7w)۳~~:$/lpWϠϓNؔ]8l"&h1߃l}Q,KyP>]s*"#R?"%^uoI9@ B0a c/ U8xf{x] Cx}Y DR `zy _Inח^cOn7Q{مZҲU.C~z`Wg:i{-:y(U/ts:M8iT]2VB|_@Wq-Kt˜~'WvaPq>uhNT _soa~\+ݽ߱a|eyS?kjw-2g>{ޒgϧ~p}SnP(2 us1ݤZgAz2(ߌhڑmȯ Ҍ@ɜY69ϲ29rdκ,F2=FO~*ӣ׹4 ȠFN#ć[9msNp0~q,zO*+&}#M6HQO 5?/m5~3[6/6Lg~$g6/ϋU!E1 !/璈PH$ e}E)I.bw#?g!1)((( /5ϥQ4:_h𹼕QEe/}0U DKW-Wv_ψޅQes#sF%bu aq}*;2/ٗssK*_xSy?:o B֌Oh}!vJ|bG0}G00dλOf ",^ Յ ḥFd`Q+\&=_TmLZrSOkX(0B`̀$IO"~VKJYqVGN@Y/&+y2t2* I 6(K YzQ-)~e*:WQЬ-jxyLFqT4p^ 61d){EG7woЃ @q`X1_5J MnUZhvzVB+nQ:4H[ED^F6yXb;"oț]U<Hƫ9 Us֛Pg{SX ʾko4m/Q}=hed^i[+mT}}GhznޕƶjiLM>jjt65_$z2=3=L &x -e>g7qD]"\݆ky[5]>mnƾ]ѨTN>tרkR}. mo.nBPJ]hYiM9#deЫpܻpЫpJOH[)&JhhԞzGI&Q&:  \^X ԰9";";";ܛ}QY;K$F .Kڐh:Qa|(mQoƎ30>盕pY+\k${P dVh#(B+ ǓE}<\Cy&2R- oJ8%V.\Gs\}p/A(t1JhU #A3Zsp`QZx RO`%:Gn8U 4i%qţ0&9eaZ揮"b5SIq>Ul뺟Ê%5c4w7#ʼn9Sfab/F{mL~k;&c [) {'li|M-|3N ;vNCo=t^0Y?x7o7ano iM <ĚSLctĈpsn"pnv?#X*B"UGU2HB+JrM3r>D3|<֞k'V$a ]n>h#:reEpF#2;V<@m 1ol]_>|g>hoK  VPd+( 4d]嶆H NIܯٯٯٯuDb]=њĠ`&8 mBbA#Xea *j /1 ȺdH')K3L`g3LCC3W`T# N8R8:C3@85ʙyq]J"cV"ЂtTY]kN_܆I4j, [G8tn ء(:}Qn]$)r"qTڛD2ND !hx(B`,2$ h:1Oz/Ϋ_v^(XHQQ514Rᨏ*sI(9P ^ VyDb9ɋ#4Q(BNȳ,#m 9t "be{[\ 'wЉ6 Ur쒎:q1TA.9NHKwѮ@UO4d!"OBna!:@jI9 - v^ gNAyc4c\Utg9 AEjÕtq|H9a!^돤 @w:6}yq$JVL- dhH(UC6H#)C5MJ*-t `T !P3h7<4W[ي Y_V6|܌M'8W}I 96j^(qp1ܕn,#mW!گnpXm .!:+oۂ'*^݌?x('vw*)m(EߊP]LnpXu"0a`r)$;DDXy YZH'%cL\x Bth(Kfyjְ D5 8 r8YZJT2dxQHsT@xa2c:08K"K3&jB Ҝ'(̹8"I԰@iX Erq0ε7`*ZppQS61ZVKc %V ȐxN"Uk5)Jm,2YF)M 127767.5T/ynFcl3`ϡg%2וA hގuejmb]HI1m djfe@.p@Z[\$#(jm 3S>d-nPOކ$2A4C-3Hua%Αh=wng34eE"\|7on;l,IQorZ岾GFwۻ7\~5X&\G?/JvJ4r1h:iX2cF+rkkaRS%fU U ;/2' \D>ͩԋ)W {[;Ѷzo{Rn"U>p6\Ǫhߜw/={9?7+m_vXu Agf'Hg W$9A}E6EIvERXm4^wFVΪ% _ʫ''fٹV6NG\j@29sx5>Y0y?F[м' cT7F#hV],@qnX=18luu;<^׌{Uwr|qVgBT2ǡipA'qhAnuƖqiiz#^0vzdc^Jjur8Ó L:o;>KJw \SEsy 1^^!20 0 8b|G WAWSSQ"z.h[͒(={O!:1YČJJs,*Naݚo(7IMJ DIw$^R)#3!_ GjM={33A=:%c|s'iP|xjl|IR"X- 1K(g* u%HcoIܙ 8C>haG.qcfCsc"(`.%-4K%B % |K$M KvA$`emI!aee!)!w寪 ഊe ( )_PRq`![D+Ψǎ.=)RQj[z$>,.G+#nBןD`c={@#EeӆLTHAHK1p?a0'zzi4HABSeK `0i]{bkbN1!s@F)[J0!I#f"dVET1Z iKKN#:%JHEY2Qž4(v91)x-;vz'LzuYlrK:[tj)0"0 <}Uwzӫ Eiꗵ/@ێr9F ]¸YhT]4-{@!pJR*a ĖrZJ]0p .*3&#-4vuY!w)J)MtI6 pJ7?X9\x/>]\ןYw WUp 5i'㟟 s;wo^i# 6Z|T:&z[XxPfkCH^ɵn2$ Iǡ[fs//M6[?jir7h4z&p4pt51}FMy\xswG>.W<՛(ߪ8]Q0!N]-yuwq+TY#FYm_5XYo$䕋hLGGmi7|p@햋A>v;_ dҚSFHw#NLĔ)^~$V#|uԉ1"o.}fq{^}_/T5olxNrq~vŃc b=*X5L¤FHܞG=:Az?jǨVF{i;Gs?Y\<68Y HbNTeT*~KOLŞgW ܣQ)vv@'aGS1X;X~|`I-;Bhn^M>Ӌ|u!@upLW_i'WOd Uk&5k:n! @ w:lk}>VV:$TÕmuXFh-1h -UuKR# lT#{RȀ^ M^p 1hlSC0G АBl|.I M!-Ʃw7]6Y2Z:.Pڀ$(J@)k   De&.C5:, w9/#հ1dbҙ.YtKAɲĬ!$䕋hLe[iu-}Fv8f/4U !!\Dd1X-&4?z\ q2ѩbp!1'i܅&W !\D{cTDl#Pȩ9"i F=ncP(}i5Mi55ʙ#Jgj{(Պ 4Wls. iiud~d)ONHj{*)[-i:QJTXLFs3F;.ANڟ}lsѪUIQ@[Q Qo3n_{E5\) /WH6JG_}*O 0b+Pʃ?ƻ+,M}.ˋOWUざ!X2D]Km{d%`>.Eqըbև'9bnf}rS`=L龄,i%F@ʒF~,x ;XoOU^X@GÖAvJLd(|4k!@;Ee@XcD!@/ŗ&W(g;Ȕj`b8DVs`IÈNQJҩdlO-Q߿ہ`PLza|g'E߷,n.O^յOf+p*oψaԆ)V z$B M7bx F{W &/v`^<3Ѯ˟/kX/QfVHuL_yuX$xvaSҾ{I>;Vc~ҭO#O "*D29Ĩnw!9韧V 90')͇ForCfJ-,%%={ܣ˷v4& yg9FXO$T6@URS9]7>b}KZZۺKZũk2ꆘQFw#=9}I I Fq0imrG fDI +^*]QxTBGqn˧(m%A:ṵZbdg pUfO#lT%Al/\ς`& bC ^r}tL8F|aBퟑ4lO)W>XTigDZD3Abƭ,PΐU\WS ̴8zayߦF![Iebԧ'ZILaTmKm%5r͑)F-,_oѬ:WGKp%~,_|{7qp}jǘ1f|;)TIB+42cK N,(v ")c,Qv׸+ߢ߮2~z;[\Do cYչo b,$xϱ<eZ!mL. KĢٙ_܄o+~pV\\_x :r閨Awvwڊ2%OuFo⋸%<.v_ۯxMunL| [<&"Gy ܇c,Ww7sCBL dɬfJjd}$" < E\LYiye@]–q{s42vt3NRG+KE|}\f[LAL8r!^>3 /n(կꗿ'%/MȎK }lR{‘,̤'ο@+',+[~+}_ #X˯nJdOF=az$_EdȾN2=8ٺ{yM:NnJG x2 ˤBT5NC`T< I8EwmH_vq+G|8Z{w_nQPx8.P~D-kE99g@sG*)Ϟ`y+%Xw>T!Y* k`9Ha0{tR/( Yʰ8r PN=qL]xG?g#fo=cG֜kmoy$b?s`8ZI5 &BE Y} 8B3/?h_\seפ,r.ŵM֠gݚD8X QB+5yvkl\=f6Ϊ{ûs%G$kr;SSuVS9m8wNX կ}Of2&#ʶЌb'#vB\B-h%d#Nȅ-Q!%e!PT*`E\?17$ :Jq-W}R\hb2D)%5rfPbbLǮЪt$IJIEWKW|?|->p/^*V=!>#gMSZBlp(F$NC26"%b&ϗځ[ɨwL8'%ҲG 2̷՚!Yz@H|5rU|V>q )B[1@;h-Oǐ4-6&m.v^v؝6w7JFB+9'>8SåWCkܝɇn|<0յM&笿]g:F _%Xb'l# C )2!֜hjD*Re)kn1fBf`/ЀM,r)00v0za)g)yݸWK,כ} )n4PN#mg[ Zsp YF*A`3S) ̜"DfD U' dSF*%o>-4XLôgƆQHhXohX‡ὠr ~7z7Zh3WìqX: Wò6Eq2iq o"I&*W:lFn]PJ0 yų^d%)Ծ{$3S6/3`PC$4HER _+Ss,#,$MFDbT*6b5os ا @Q{,Ov<'+^vqH1q8rOT}."'s]'95ؘ֪&,[m{4~ҳZnvOuJ.xb !OO*kV Ʌ `='N* D0 }!F§X&&2vKy HXC$D2ɢ0i<%EQmQgq,>7Պ8ҜqlIU0TSec 3DOAi#  2Ʊ0/9b bت(C PK$*QKqFߝyVD>gE̩=}g#rF﬒|m;ΕR'L]"c:ŷ79 0X ~˛[3&VL^^]6[1^uD+t{`znP:YֳWĻ}皋ǃreC¨?hn2KEgፆm'pkqŕѻWւkv*0b(&ZvD<(iPz絪kS'Ǻ5VǓgU{mNHb0X*REç0RsیIJWoj Ԕ-N܉E/2stC%H r&!&M2 Zذ&-GWgW3Hnx UBGAh񓇆ਟ<5cl c/,VcN *y.yS+ QӰ չmj07ƌ O*4ϚԵln#ù><^ #Db  (Is+C#hѾ?a.[(D$Z ̡]@Wmhyudچ(&dFP!Mˆ)lJ8^U>ɣ흻lP0)fHV?yh0S-W8cH4(Dscb2=sm-k#,ZR[ݭ:bn=9Bhq_"CB}>P\tx}w9$IR vZߋ>@8k z4Ά´o l]lC@JLɑDWoN(n!sAkRIK®8sr8SzdQgq8HVJ k,MPg_ײs|V;-k_T>l9 |3s]Kwu*LֻW`rKr[[IosLլ9p$嚨K _2l1Q ov<ýgǜ~u{~he`4`?i/oήN`)Lo`9ʠe1YTH$c~ f5޸0̨[AOыgǡ+B%i}Ѣ".i Qtn&cKԋ&rUZG.)yw7h9B*+#S9kȍ ("1D(Ժ+i&l7&ΎvK.1 X_fU`Jv?VUz#R䦄brAr xxH蘪 &?3s5/f&z=&h_MTT ġ=ve׮ʶM4:LZ{@onB )lhOt~54A *~C/A 0m]F"b e,A?3/cA5.Ą"FGSX6b""T%LGr-+sn2K-BXˈ2c% (T#8ᚂS*4&*:r.{KB5 E_7Ij#0! K*Dx$>k8Mɘ:LZXC"FF+p" a &,p vD%H<PjlFy,j fi&L)aN01EZK+e*P k*w#s2I& KP^Lf@ޫF^-^<d>l [_4 ?߸@DA]N4aSXoB}Q+ѕxJ~̗xw!aYJ>JxLa$(Mt&MAd!R(n<2_oht %U2݌c;}5qI7cf}k4AR0øh+kJL 2ڠ'2%2Q%ČҘ^+Fcjc$gBNc@O:m:(+Aظ־ʸ־3Jp ^5Fc^_C/rEuȡP;KB_&l7FA$\ 9gPɀL =+VUq˻1ڹVbWwF+941EΪp*\  +^/R.ʍ1<!NCH咚+JrRҘrEu^oW2p- üSơdoa"X5JD΍b8Kp-5M-PDK$:\DS`AB@&qU۷}]g"oōYkjd8xOo 'bwTsO1ipF84`Zq@k >,B^&J**$SVt,CRQs- SQT)%C VuRf,(#2AQ(ߏ cGVUqνRI5U19X- 8PH0A ;CVB,yՓ* Nͪߪ@cmakTjpF[n50ɇM>haD>Xo)SB2 (:>֊Hƪ$Td+geT1X i.aCoM' K[keM5*iq*@͑Sr ~}• ;Ih u%[AX9Z~,:J 8,E2 L:WAo8p.:$DhBT2z ,x'85CGp5>}hV0Dw ֶJ|W"X:J7],a,D&ı%49ҸppHi# Mݻ^$.ii"ht Z0ϔ([ zPƠ SIгY)F֞ojq 2E|-8"s8\ (p*Q-C]C^%SܢOAV}vpwwqCz1",Cgvfcz-,|`>~]NA<[7?:˪$Jm|F_u}l-;}[99pK|uuyi3=稡LPMZFε@ǂg+;q9М>K3x6'AB%3.a܂3to2F G73"\bM!iJrY[~ ZҨݦq©zT{P{Kş !ڟ٪U{>S:N6iMա 6%~wr?VH}7%fonڡLu)y_dr軐/2|UV溜T:^\뇁}q!'v6t=W Ҽ:8^uA@py݁Nm㏞_(Fjb}*DFvUxL9](t|8cYwg>N CǭF\-YetF!\r]^O􏹘ֈ5,T!g>bWƽtR }}ؚqz~m~=})ǣCج_-p;[ػq\hp\mF#q4(cpWXn>~ȕvh!8yK2o_黫Knwཨлk:{|]:_B}A+ܔd4Y QDe㨴ʨ(D$)PUsq@$q-|)$Rըj 79'HkXX윘1JbqBg$j<|v*uj6K>ޚߨsh?n~h;qĐ+&xu=::.+}.+}==8Bf%U>e=<=:ZFO,u9C2NÁ)&RhIMX!%)`C:Ykj*&oGz3AI=`Xpjq6Y|pV '"懳ð.J|a5i5|2sgmxZh?ի\}5{^D֋sى7 C(Euj:՛]s֟iwcUKK5SCKKV*4p";AJV؄cLL'X"#Z(\NBlE 0egkT,gؑ/$a@V*\~qRӇ% `-7DLFuqڕ'Z5Josuz#>{u z ΢ .8-풜/li|Zx*m3W~FIrf;!/iy㵐N͔!/x#%m /{rF<]]!/٦hv6Z=^&Z/B{?F<Ʃxf5d6IYn<8b7PsNFn?}9%ixeBќۜ!І~Ӽw [EPQ1|ă),a#F),cЌGa` J,),z)dTn Na҃),+ᣦPi)ƻN}w:ӣϊ:`S˩bFn|:$}xzt]Wa'_Ō><=rr(WtZdͨ4$$+BfPH=xeԃAGz2(8{9:V ~MH:_QzB^&JwpF4(8=/z'@p8ҨV`"o\ߓd ξ |*;SxX 15o8qIFzkB"DoZmr)qv5pƳ~ʱxI3Y}O2qr9|_C7&_=xD2%p۹нϳf)CL7&4*(,-eBD,@ <a4 ] JnǏ(t ѹ˷JyɾR d&iz{=uJ,ٻ| %we>O!l%sf$[35N/ӯdzpo祔6܈pr+YpGScxbG[h4:l{Ad؋Էslg%v\~ʥlVe,%cxls KYBԎw|AI)QFS_=j_f/ǣeOsT-ToOrQ8(Prpra}̜q_ap{4 ~[etkh hK G\Y*`֛ScvKTPxQaxW4-Ȋ)~d'P|q (fzAԿuC4q3Éa.d=,]~iǧh6 M~ƋY/6^qNugQ0fOA)]$>\ $(E:1G!waܟ'|;Z˛:})=qE~P/9[/r2T*eYV/#~z=-矅>/JSk46r'h*YkB$B'D!r?7]XoS=Ek W+JC AiZ5R)2HxíVOؘ* BBYȿKB1BAiƤJE`@|\JK{ ҟ~Lf&FgR~|݈%},NK 8jth$wAG-Aք ΁Qjٻ6r$+ lRÄ환ؙ펾=uW[Ӻ5Iݽ1 J)X"l*$@"aMbCFNX/N+8c3@O534 AjQřbyxL]g㼬:Ɖ,ikh>$Z ).4"F@+M1*='lJ4t X'MZ# ajN*QPf4H/^w^eGYk\דeeR퓫%$$-hr|YOHyUMZewzjk8ogngŤLsxFVfV綃հ4P]O$\Tq1;/c~º>HRH!5zxlr0nMהp1ůM[}G-4#V#Z$t[FK\1&ϿvK8kw8YݻKT =磿]_o.>\\/>׳-׋𣽼';d9@ߋ\?_4c0VgڜuH7R\#KD>34n&NnF?,eIcÛh|.~X1,Z]ר72/~I[#n/1JόJdtXӃbQ#euGte<<68+e%^iSL7vl %և4lkͣSFf!Qγy=<1X{xI BytH6b睫ygpk=3؋fe㬫{e6w6!cyփX>ePuLOWw㤙%SRεrIƤҒEH`N*)5N|p#D2ÈR-\m0q'nQ#aE ^A92)lc$*Zʶ/pʼb7ީu@9RX(A5?tIݪlȣo7q6^|h&._3{&'CmH>[]и+wբf>63v7?|JLp@kip큜Z$tQo7J)=&L5\o?u(2 ]4DBݢVω`mm qpY!u@u{~@B|i~*>%P7vIqR&N\G描PC*L3>&ߍ\0]0=)} P[->hxzp2׋ArFwUh"·ƭ-d<{J|e5ОnݛEAVN~fP+prKW?ykܧ- =tc-3A Ah+=_]pn}Ki*ZwOSKt1䎗rbMSҬ-X_>:iK[6o !ɧ WT+;].4P0]~CqE缙ۓN*A~.8i*oݜr7ĎgV GnߦYj(X{gƨzlEtH4$*d|VH"5K1"$kE,-%3X3QC/C̦c\D҆w|Iq5k}ظ='Kg%=C͗C :J~kй*ۈ,ҕSN~3j1sr[s!'Ead{?BUW'_ Ѻ\m(^U' ג?yѤh8y-80Bt 9xQ{N6r)rhKޫU=镝:h%R ݇ Hcvbq8k'MLR`[%}Z#(d[0QXL ԧ.IH. Jg"uLJɅ\xV#|BWֹw;wNZӌ44E$|"&:>ł<2I跋G{&bZ(Cezp>ynl^'"|,DFӫO.7_zeh::NuLJڬ 'B %xn08!+}f޵ה m 6 }ͧ9mƛhѲݎn&?g|pkY);^W½ZUֱ>۽fWi!$Z}ʅn&`֜9_;w`^Y'_CoɷoϋݴUPQ ^=z!*:lsyWw.c6bqцˢ e~βhCåEbUURic7ߌݿr)mxbFiYMt_{ IsѸHZOzy2( 0*F܎J 0ʀbB9FT=[jh!B_˧{7]\_L?p;ĩ+cb&פ {ď0n0Pٟ濼,dqmt戡 )g8T3iBXRӌ`MwH# sSol;;H8a.Y,hp3ףe7eXlY^৏qO1Ս]ynAw߽2MDdtwڳPP}|4\E;TCY7ټno֭+%m̩aY-n}hȿ>)AtJsIl]CBodGţ]Gc1Rx'Ib 846@=%콮Hkw> W)\pE%(oBQ'`1nVьv. c ͑΂\ LZc!7dq\e@ AMJGAG GX'#1B[nK8A "\Y,9GWC4zbK)uLKʤo}ck>ϻp08ZJ E\I{N,E|)q :1=wDK/i{dL&g|p*EnA[#31|%6àdY'rA!`p 1=ҢLoX~Ɓ̕j: xUuӠ xjko 0X$x؍g eO˄Ϣ7%|6_RCg}䬢qk}C r|q?r iԶ(33K2 逇,v4uvͣz:юH ('O_^hy4 Yp'Lw"O.|T-@cyWSTMԫ/XW "UG*-9Pܻ.!qڑIÈǶKzpbW"dFqbt00T]t0*ӊrEe89͕yђ6q(cs!=ܴBVN_MMY̱) xd2{'] I ' 8%gXvP gXF69}m ^pF'!@""   9;p6fڦ1]șY^&.~(IB,ưF?:ARnXp$xʵfTrHI@ܵ}!h'ֹկc//8*A;Ehm:㷪85qN$w>`YRJ\I R;Rz$*@9ϣJFX4CȨ#PME4!I`gq(W=U.\6nb5|LNڒ1#c[jT @&e3:` #\CttQ{9>c*%/i]\ KOLd"!qڤл1Ec끭k1@yp4ҋYE 8Xcq<Ϥ  g>b_}BF Z /.ݾϷ_&Ģd5%P}m1.Ŋt͑_5SVͅwg `.5WL鼙7p gcB˾R?37ghQ'ْo}Gs!"÷GWA+ .t|܈|Ҝ % ltۻkl*wkMs=& 1dRov?~]=kDQ|$hfx0'T$,XerFº{e:jekM'OuGW/r].{gQ_ūŀV'7?|g#;^|BCB@NUqp:Qn !a7$c'4.@o&7`a㑌y=׹#{DhS sSJj.&@jJgMޗ2L'Lb̳j ʥDa+>;[*8}g Mus'a9|ΰIDj]נ'LڤoLIR}s6g A5v-IМ*JsRv \^YR:o4y.COGX{C S D 1 Ry2H yaA5S` *nD"0*hBp})J¸ nGi;)(>&ߖk&DDR\^D݃vaٻ6ne{W! rҴ(NIz .ɍؒ"y4~zݕVaKrlֳ(koF#<} ˕UKkApD/! |ZqA[rt u(t?^<5˷gG?{r~3_ ?}?2}{`}o={x{KxG ߌN]g_ׯ^|,ONF7xN߻:˟'/7_?闟<g=3Ex8^nz#VD_{ϞƁp OL:wv=5SLmۯyof%` 4i՗=;-8 _븳W7i~`hFO?M/=o[)Y,/ni5%۝|Ng.2P5ΟAy܋% Hk& Vh<9k)rۇO=amw^,،䦺4!Ξ`^_}93w2oثk?~@;|rY8_а\O/ Տ}@7![.|}b&]&=6ߕ}}@Ѿ=zЛGw<}|~> $/Kp*2)\! m+OǠ hp3= 9wӸL|A#Pz5Xݗgjr,pK+G׃We73Wc_~nߧ0*>70ի\^낑Yw:x54Ge=dsc}yGӧBE6ɪzcj邧 írQݕ7ѝwR8'5\"JC \?*2`( #6N8d7»*!oNs4ޜ# i|sYy'.kp|Y@/A ͛|0BGmeD1Zό"FJa?g{3.gNCOZ鵩0U! dB@~> JIf3&1%\f8e0ĘPQNp%5rnx sKjEpn 0C9(ADW&]&;@;aBq(c9|Tǁ&ep!0B{I8x ą1*LߊA3Bi ! ewkJ fm+ybEXMH4r{N;bHQ+C1lZ33Rhfw|t?N5m&kP,XS0Kf=lV}%ö&88rުő1_t f>>OflGwZ˺z>ܐQ?kő4\= i2 % f<&@i?#?L˝_{M"݄Z}ID}<4[.wpLlOs3K nFdp3L' ɰ{~ Jju~iOȷaݾoaZ.: KҁȢt<ŹtLӰ nlke0H B(aiFEM2EXҁj+<%\M}]f{^b*= &rg%ऑC6Rrj ;hg֊HV2td6iɨ\X?iZ 'ZhqΙFV(B@UYA/S[[ȼ};D^zYU:1W S8=2|b{Wsw#N1s. X dȜUFCL"@V(K{&B'$U Pc`)3A@Ic Ј쾡Qo[Ju %(!)@qS(ː4It8%H$zL4zc WF b4=2\sI$1iˈ:AU&S+ e66L*]sY<|r<7' Bl|'.e-0Bh ڏ2ޡR3F*M[L(gEa:6YZqkl g#{I X\~}AgM=._Jž V̞uHO]eiz;+kD d UML檚^Ǭ`uS9ɩ=0cok I,پ 5fn 쇴΄jv"ۓ Q\=˔i]QJ S ;xpFiIE1 KYB2͝t*s%kژֶFVni6sQ,/\%6xs:^aЖx Y@Rk+ޑzo7E(,۰Pw\>@7ﶶ)Rn+@GĚ[% Lyo XiI/Ę&Wq>[!.GDƻQ[ߘC{JVT%L1Jcf|ܧJlip7Ң H鏗9N C]M XP]4YҸ[Jif0k, zG٤j#'8Tëִtbۛ(%~QOP\?w%1W-Hh‘b<:sIbGFw4Q^F{ǃ\"R;jljFEOgYK0WMuL3 79Z%4W^Юxm%5k?] Z1NP=њc.qKim-:Jnt֡ |S*3H m;-K{C--[l<@A)[1=NpH1%[?tZ`m.5m>ieC.VLF}: r%v|udנAB^m{܂|Hy 1>_JZHaF,e$'9Ŭ_6fh %}J}@R~(ά)9_>~c&at`=%@*'1&$kژgu PNL#+WaHﲰu`})2:زv]l.@i${O6& 3ѰNnYI/jxT[<{}7F1Uǫ۪ +m޿nS.EW,έ5Z9 3Z$&+e"uK^Pm emm|{f)\xrheLYLJJR3-M32"˼UR! e.U:W_ <'hg%(U;H( [%SZEj))0+>5LqQХjC}U6lAS cnuʉfFHGk`2Cdr1G.&Y78T5. #MX+xX8Giʙp0d53f`!gH҆HDMܤ?VDEkpzPk18.D*]뻫?Fe_\};9/gKb~M䊗TYr\ap;k}p'x:ٖR2rj%x%oR}TX{,M03_uu?'7]Yqe>3Q GEZS e @&T9R NUY[mF&wonyS j,Zf9,E2֡g+mÖmm2 ˘ /2h''^[1W|:`7ڿu-X]1BE}|.ywsftQ?[Fvet#F2\<ݹ­FU0yV{neanQ%v^;7daQu(=BVnscgH x`$LCh֛/뜴wOF 6`n{&oy/z5 m칑4@}^}ќ8 \Tp>;H{H4  G9Q svfTHhgZC9#2yG] +iaTp"dFHFlDrx~QBVk=Vs*8QLr0ZGcL֣k38]wc0 [Z.0 TFa-yN5J֐ˢ* h7ݸ U-b'ukm֑%1ْ&toJOԬ14rѵdHe=EJznyc]1]`c4)w)y&Ry ,< KJIeYdyVQ9* m&,i3*cPhAYΪ "HS3=Stb>cGg\tICʴVmsإQ˻44~cQU4ʁ2PQg9ɥe*dS#A!2s*S0;:0Z*utvtX͛G}fR1Y߼YsysmdFQrMH:k6:ʯ#Dk8\ nLx|N554'UّÈSOGLv=bb[̿}VL(&ء/ƃ\/VQE:7k,ѻWvU>YVOWxw.n}M5r'1/eS4gʬjo,ʒ'LsR:-L]6i U12/L3D rikVJsc4ndknd E(8izvm:'jz*bzI_Uo~s*GZ{zUԾN#b 2X |uZHF)U\)zl;%F) (0ȝ{g>^.g }rCJYjx7M~<[t%͌Wwʹ*O4fT,~YeJ$uo*e4|6Ʋ\'-14,kl# o r% :_{x9/˒1! [H|*FFYT *<+oJhPL[ƙ lqHP([#㴉i6Sv9I .G~\m'{>NUb^2A5D!JePUJ&ZCMRIf®%D;ݔ8ټ8$It⧩ R#Yǧ{"_,Tߐdg1٣{5 ߸|xo1R'f۳_.c-[ ™εJ==x4CyWI$F̣: J\'f*xPqbqF-7]WLd1~-X$+wwR}|TB@y‹^.Xu}Uy<&YAӋ!;\"/I: ^}+pf֟oY_nݏNd0TU.h-_կ#qJa7m*P ؛ c1p )ѩz.TbL%NwIfu4=-ۄ*HU7…^ã 8zo㣫Fl-:J :}vyJH ӥO2A2lk^-[d+aǞtWP4h˜`J@?*q)1ɁF ́Aee0࿵qΙn,~+9+_Gtd"Vgu"Jc''B/DRW͟Z?ko#'F/*37zlaA1NRClP4zŸcܑk<a#xqƝ Y.>\L[̩u;=HPei=|PgB)H,Tp'"*xA(!L4k!7j{`P#(No]#R\ /S/y|pmc @U4lr[F T_4/Bת&X2i+ٔ]$"TR[T4j8<bken!2 B<91ƺ1 gU_KFΈ;2~.-6fƋ*n(#DŽ򧏔R14g͂¨#'x:CČu~4oXE%On&^HY.WLw3"A-(f q~ތ?`PslQFI# baI_y!i ie&J,foI>Y^,2JhtR28әEbiZt.]t&p:)횫{'-Rp$TdErȄL DFk'1Pp@=AMB9f^Ǩ+PHu! ˄"`F':Y"upEPH)_ŕW>= }xޚ9ؤO'q59m+5/BfVJki55Ф]9bZ>^Qj 7Z!CUN2m _}41% t˗)Vʒyl䤱nJg7_~qTHB.uX0|Y=$@ѻi,ݗǼE I i-64JNٳ5Rq RhZS`XUd/_9 ?DK#.gU1jӍ}wu| 4  =*r6p.*$%'Rfe:i1 9Zn=|5_SU #V }1EbQLBSpC~NVfg`E;2(3!Gf~:O f} nG=T|Gl_\m1m3Di3UywX;+/RH$$b( JՀ.C>ZG h-F;EX0Lu 0ރ,{ΘbA,S>_[K 3qRV,x1a%x2!D5J.6&:;U{RЀ߹Օ+Ø܀> 7\C>qfI"jpRP@CF/:JB0>ן#Y-`Fja!xhM`BՇBOE`m:r ֐\`AVBo,>} ]My+:}1M[ M_4 /tFNEE d7UN".ˏ;\ߍޒSGJ*d-پ?)ʬdZO~/FJE.Geo!z̗ӕGR;CJ.-4eq%tq"hZ$%*-kJCn$BJQp׷3ˬ)SHi$ LHiY@ApbtywbcY*RQY`.vox2쬯heBvYa='*YW6'/eEes_<5[6z˘,_5Kx=P7.Y 8yL[( Cs}y-yءw%G3\_!xq'y`#Oxbsgw>1p(~=F,WTt9"~qo69L؍6D0.[7DoE]#QBobCP$pChL9XS=D'Scb}OL"t΍^ \,j5\=UPWWX?jnV_/6rUU$XYAX J?1(Y7d]"-T M0ј@Eh0IsD"Sӊ"Q^/kߣ{q0өhU:1?{WF /3%> ѱ=юv/@9H-I흘 R(:HtUߗH$DL&=,O`\Zn:k7 ~L?v㇫1Cpv؏ 8Ct23XGє /'1XwCrqye 8zp['`r8Ɖ¢#i?0̤3dR`RD.s!aR`2PX 2y@Zpp`*A@fqEOw~O2m8~Yl{GUpK~y I$$ ?C./<0}uTeѯ->< ,2I _~}0L$} 8C;f2®3ZpGI۽hƮB3=染 75j?\mboփrRҊOyÆUgܜ cK9u p[u 1McRyVYni$5fd0i4}aQ H{q:]*PciLؼay(P⾴0\u$v\N \oM mP5Iqm+N񒳒6)(P=}䊶IX "ȒN@ Air $t72OVhC!.Kcy&71ha8`<Gpoן~_/=Orm@RNEH֥{ wq,T"(炙 (^x5uq_}lbǃ$}흒m~~ t 4c!?v%m2$KR'hD8_ؾs+?V\pqV8Jz`q]NP( SQxyLaZىI2LOOw>Tt[F"Xv^@z·v1gSyd*rUG7pMHb t:hL/A"(}/T B2\ls7Dϱ|K7ƗQMΫjNdYosvxMjiDYαα+Lb:,[Ur|ʡS,l=V[ sQ>.&A}o~ƨ XkI~VN%m.SK:Ӌ^Rɻ31Kh¥ui*<9ۄEFYH1uR,n;6-2tM@ų*nDUD5laK=p$T%?u:#45Ka?H+u54DŽckp-&bN5[N,>7vw%d!ȳ=ڽ;sBLx +̨h'~5kĔkÎa"<%+OD${LX߃f#o&_MSA6zu}];Mr޷`t0Dƌ{Vv{[ETέ⹒G`VP7" wǧ½{ lwSso 8ˊVTWX#u0LupӘf<)<؄o{"L&xHro/ W#jމ3[k"Q;. Yn&~ѽzDse"H]BkԜỤkկ k؎^5AKN^(:S,kw(S%E6ׅӚp("ufpz\Q4_5_bݹss0A CoGۿ \L3!6FB(b )E-Jw3}6TC`!T7h`U&6xrXkPD(5a8͘.Y$`7Ī|}SҒ.(xtpq%sq!~ )4OB /t?q*6hQ^TZ9aDti,ؔ(f [HЪˁq`ՊQ^A\" 1E[:/,}xXeTCpN%i+,KŸ,eUWҺ0 /z^e0wZgxų?K(]]F ^? / <~a4|pEø_R M^sl1rN ~ĸ͘*dĥTƹSřMǷ59mH' Mހ`Bٙ:!QG8i /лH%p1gĵ?:IOH~jU>q o?RJ㇚}s |CB.~h:BEcF}Ɓ8[H $m|J8#eWXHAq Bc!!rbpv5 ,gtlg3&뮽BF&zkШiO22N. 8'xW\|F(cӜyk.hNrc _ Rd&{s)T#C؟w3_G}n &=XL `}7ٛ~3\c&exi!erJBD:}'L mR|cp}M7*VOhA[oUR&,0^%K8'2S[5Fh*\Ԃ"PyДkAtCNXPtR2%XX_4vs\*4Qe1u˃Okؾ꠵1+'*?@϶ PHpEhcB$H2,aDGDpI4sTAxꃻ Ս0w\wfZ y T ?{>*43Y ^:4WЇC .;Z]2Y?4ză`dzi" ?,0up@SL ?VW. [6 ;fKC nԬ@f~L8 Ȫtv /F^̻Ǐ_( :ֱ&mO=djb(hj*#?B'5}a81.X3 ʼ`'T U]jx> ;+C$̥1{\}SDY:k=ǣ2M;:x}"/wTTvq3-*1h h`@2=D;0rᱟcf:(ŌLHE~ @m*-(?Lܗܦ=s5i}'e;0a_?:"K?%ie/'_?~ae_ޱ__xr7߼z3tA:tǟAݤs1~˅ͳ__O^ u|=qOad_^$`ۋ$"?ɵyo&s(}RfCB3=vn^1;ǯQ^pN뇗w{}+@'N+Y _RlO't"UQZ$*EJ(I S51(lY;3dU7!-_FnO}dы$vtoo\zt~b t:]jϱ/{ܶǿT#S s< atA/O,?A?޼LO <{tݰ${ HSx{ LsA~W'Nfʳ t_ްLΚ۟In2-]>C^ǯ?A1_O0SG>qO}=PK8Vn$Eh/ô9$KP;I(pG<¬#l^LP[ D@J",6:18vkH0AgS~ms1WtzqֈK֢~/Vz Fq+mdwœrLX(m- $gAzvI,#U8gaH󵰘TҒN|M-yDV.yܛx!kSst_mzg,,TyYZR*KovF13Fq[h_&Hq@ˆN;-0J % iA9&a'oP!4G9ﻷW 5{{b*O 9s Knet+?hEZöepǕ2Hn֤f.R@_XR;fmD8kXYlr譢-Ab32TUTR+8`*DVo |ƅOW!Es{ xdZOZϻ ^vIxGǭ_dpHМoSvαyv9UӜۆ-KX *BDC}%evκr 3v7o TE7abU_AǟĔÚ K1V΃2KxI O6n[߀Yɸeew>l*ٜ@fͽ*_cЋVF+޽b_]Ct\kdj"Ve olL^L~74 g1Teiy.rR%M#!}0$hwn%e1x!sEaA iDž>ޣ-h]hU:j(L7WNJl4_&1 Q"T:^/aD/7}Mu1FCgṲQس:RX]\ӗo_;0,nX˜%-0^8&zЪ&+vA"_ l|Y<^z[l 7;|(=)ݓPLߕs^ ZWU݌N`jt֡Mc܆İ؈ 4Bq `RL>lb3Qr}،D+l!7`ulv=½n*%wgUj& ^S (E?awcSJW3~TJ.HH},wp$v_*ׄ1Z]Mì LyL§.Kx4 =1xB)؃-4f .ݟ/u'^pt}#8ܾj4LVӕ.x+ѳVTLOEOt@q> 2pFH唐{71E}]8xԇ^̩V|`b{ْM̤/vg/mBD:'Q*6%!a߷~^thM80Wh[_| 6jZ>3yO2Pn5@7u 7]/tMQ ` >4Hűf8Jr&Cn0fGX x) Ihp81z GTܲom7Y!ւI|]|8AfkpJ8kamFU<. 18c$1Z/89 f1A^bd1EH:&Z KQHGNkN# bMPPxX H@)ݰh>g,v,1VМެ5)v옰uƵI܎InQ CX=gn`D $1:߸I>=8B# 8ֈ"BX LNl̄b`yTe޵3B=|]5`P`NNJ32djh4,8i3V0#G iC5H5iLuv"n+._シ\hpz6݊/T] BxFdޅ95~8 KxL|& Fr3qNʑRr#,-6 7w[)K0zFK N9zcX$l/=Z3F^ 3AUN~)LTxYsKk{sug (1A&aZrgm' ;M0veplOߥID/81iK%H ySFLZQqI2EM=`;uc&I-Q[\׃:78Ap@7CzńcңSסSCcnBC~ }+Zu` BDCAjJNR-VQڝM J;>\IBRJEHN #ƸV1.)y|@b нh1)lM{8}HLj"r|yN z,o`HRyp;F A&}n2Il=vctzx!Xc#c9!G3H5 %4T14P .6e_A"R`Lq%֋>D*$M$B#@ad\$mK8כ+F/"+Ty>]k<.իNyLPTH4@'_t,F.);}ro!ߞZE7*B$w!o 8cl./κpYۺ~}\GP (&GR#T;է,eZb0]pƐ5J Պx3is ɘJg+8JH)}ȧnJ2;J:ti-`C_HF p ?鼨X#DɩcA3zzR3)1(ˏf8J9q`1J$v}W}ԜIfk0m][rF9h!0l&P15>$}'9*Idh>CܬՖDRpE>TJEeA{U(OVگ$.bD3fDb8 q`T]&)L%D`_3u%e]hRu^f35)A>O15/3._+:F9.H$U3`kLQ̰-FqT0^UhpK,ԁ[TٿqZNP_r~NN"YY=-ٝ #}ّ` 3C;`2o:Ov*vWoQ*c{sOwhX݇,r9<[!'Bp%}iԓאQ <+: d 2$c&JԅR^{,5\TNM֤)p8 u/߯c|1!={qSr'#k [/Uf/>S.{45UTQvF/c#{Z[ΰk/\~\ݝ AyK ơ+W8eJR鲲VntYeWUY?f8%kݛM;SI&Qhw]o#߬$n!Rq!rEg6Œ(I$#"FQJ]ްȜ'pl25:#,(RX(1̙" B>+v Zv( NuٲW%NCZҎFp̐B)eœPv;NVK4?C2$C-n8qSSSHq>ȗΉ=g䭟*}t'IIq!'FKQN|V |D`G;qz}IP[]݉3P)K1\.%1ıtN9还c NL%Şώ:9c=!~jC8,:;ASu[#~zқ%:{8٠):.|LE-gR^=r#UԆG*.&h4ח`k }9t/0+J12Ib]o{p__|Bs]s*: dPE'E)ګT` R?]n=6!^_̀1;|86gRo~q2?6?ݭ Үy[ԘTZ.P\χ#kWDج]nkI;jkF9)) !B6AX8jh,'WT1N:?ysFkS|Or>U|/cb8tqD`@ 0r]]"XOS7ĐJK@b[WHdpC(!4tW:FNmEh0jH$\ &GBZ cTNe"9!Iv]#$SBb>+5QS=Pѫ@юK5Ks]yXmA i8tG #ΙDqp.Q6 #H!X§1&T%(#9f<!"68q%AsJ$q, :Q@{b~D|Z=+I `{ͱY`hbS\ooM 'dyv<8ɦpۋ/v\m^}eK^|V]v P0xܥ ٙ0ybo/l/.Pb]aY9͖wS%_H[RL9-TYH)a&X(Q  ~q+nJYq1,6ax{O6WS e N+Jq1M5u1mkw_r#t9-Ղ0殱݊]?+&FDO1u@"=? )TɚfUv!\E=tcYY7NuEu|ĺ dfݼjݺАW^:%ƿwۓZ0Myr&KMaXb}^kB3ARg:Ō#D4:1&8f2l9љؤVD!RTrDz6ܼ7zLq !4I5ú0BP",OX/|/( sD!3[M:c ,/%Wj8}cp44+GQf5Fl)/_;_׹QcH2p!Z+~{KS(}y轎 1+8ts{D&KYؽpjj1mK FhiVZ_JaK幍ŭjo?cFz é@YY ֋>sEl0@1\0*a}GX85QDLT,7*,˙0Aq%)Gj($^ W~yh!RJ !:@*0R+v׷r2,Q]|R)~q磅U`YfM]7htn V">A^dh2A5G1;jy1w{櫧*/z7npRA!hrMhp$;7MHIX?a¤&7zB&ȭw~'gSU'gî~B~d!Sڹr=O۱v̓@_hupc%;F=8dir|bUfUϥD\]oA+gÔ!`~6p HB0}}qysPq0)mv8|8 t[AV@N}* ӽOZNe{gNq`9G pck]V]U$Hg]E'ݖX+o1^?oQeKA5ExV- go"SO'|z b:!:M2AIF4FaSzO3ڐ D3d1)gd/;kߠ?Lnlu%n/z3TcUj$lǬ~6Պ/rL]ZӒD`S,7;ەVv.վN#FbX))h)1IyJ3bBa=D25X,SI'f.A=̑|'Ok[w 3契65nWLd!(N9 ؂ƺX {OHN 0yd K]*Nl29g)|MIBHF.Sgn RT&7x'3aTcrSۃ%W$1e L7U+%jJIVUaoqkX7`';N꺕~2^T E֧6z!-M}ὁEd~YTB %uIۅxO~HJQ#X)1NѵʐJʇe$T3vw+/8r׀/pkSu ~Lg-W!5qA23ca晘dz[gzdwrXtv'.<9&(*]\rSv2[nV.>kSɆf,ṕrXWI/@NN/׫'{42hQ>x#5oLs239Z-̮Kz _9qh--B_$RElfC2_ʅRa6J^}sҕ/x U+Z[qhh݁lLXlB,S ,WN8@T*Uq&g_ihyt5VGiB\jڳ:-WmboRݝ>v,'բ(_њreՖ/Jtxi nvtlfav{4wjdhK`cEMג]iO߿=l xPYؕ1Gg};6W$.)փM>vL0r &;79hIm]@Ff+-=ߵl2zK\vܐbHc_g\f<ߎFs4V`K5WP x`v f/[Apeu)CP'$afxxd~Ns%W8Ay=2S]g1#r72~/X:ٟ -ys5מce܈+o#vG5bȕ=Y& 9v ? з}أ{pØ#?;Ђʏ^ NGS7r^zV v-ab*ܝ£=IC!y-x/@*oF;۶Gn!/K)s7ɑ֜^Ofן꿗pXz#DM2^@zD1%dVCj7tvtN.'e N꽂$ƲvJ.n}N'K?-8%ꦡ"ѹ/V&^.sH&iTVY"x3CvߏuKё07OusGxQ(} R r Gg'ZT\9ͣQh1jRZ'j/eHUL^̟􋌱9mjSeɱ$ZeMZ;EFSfكZEB d @gsqm$)u.y^.o+|!]ܝڻzo4FC1j~oq}@i? >|^@cM}ofSRs.b+ƫN~/h=~^~\NJd?}{)&fX2˵—wpܔzΏ&WTj݊KYjC1 (dW~q|v=IǕGC^Wi$RgV$TȆzhdb y.ҹx,-Szi5aX 'OQwXPvd+V:tf9tX!t] 2'yy a_ ic`O5Zj2PZI #e$1#drKHZ6+P 0ZVK-#?֯%ʸHNŊɬtЌHC% 9#W,X)b h:{uf6`mC\t@:q.A/Bb  dS1XOs(ItF쫺9zMƛ$3F_Z8KrK0vGm N z1QbIZG WL&Αh9rqe2Ӈ&eG,!#7]:,N1 .h9)fDoiwVGo|aŹxc!x] 0.Ӑu@Hm,; W( Oqgw>K2G?ŝW~=*;|U?>?=e:W'3 7ӐI/dL'~r~7RVV1X_ *=x/,^QI.,A ٹ]PąKY'urN1 ev0`c&AU$rF!g~S*9ȑ%~܋VeR%3XEpHMF¦H $"0#x %fKFruG:@qo(%F Ok134 !AҪT:r$ɞ YG\N!m)^+)>̦ qB,Y'ŝA$0I!iƄݠ3zNx>hRyv4Kb&7H-lqbkb''ɾ LXôiF8e U4-r"|kI^SWW+MCF˒x9ITMe)j}AҨ1mNjiNER#5i:\p`Hid!,_ob:c B1%ÿ> ]A B6|*]55Y )y+I>0] ReG.@\%jA>+FgB2ȤZ-5 42^Rekyq7v{lb⎓-~V!#MfѱFbm{ZkWaR4aś >MҤjꞢt*Ykǚ*2"k.h!5relVIe<|rl1ӢCD¢mHJg);տC>ԟ̓1^!h'd|'D\ h-l Y8KK|ˠ|{8ֲw?j{4ނ\.fN~w&Dߗmˤ`M{CiH)kq`DC Gѯ/vD˝ fDsv,VIw_8&}KI@T /.uمOD}L]LV>^>~\: T]H'}yʻJ)9Se4_}ߗ pF[srgtb2דWe ǿP4]R+*W*vd :zRCz.l=( Akv^rpb(6.DeR2,Ř块w'ODn|Ǚm /n_x+[G.(:­dNp챀!%"EV3:o yevA؛|;ۅ9ߨUI}J7*Ay[ƷjW b jcȅ:BT`jYbau4\#[Ls]XLP9@)g 'Petܿ RbFN{AA5gnMY0ۃdzF +LWº<]3h0߰2.:16Ť"4 r\mnf*t=TM!p.70iR\ ;{P_!/{d22w㇛T,GfB{68lӣ3XN)h4&q[H;^8ɜݒwKE}#BY$/bbgtKٸXz ).&q̛y0O hE~ ߟZ‹O'&iu\ZL"x}J0LkqU8mFN Ѵ eCV`Fa֦ªo:/|pGyqxHF*UX u441Zp1u P{OeSrBf*}y<\:Q[&^W@H*^t}.WeDZ7nziwXVhZ5Y7WxZUx#jmEXH.yٌ\ԽsTn6+A@99;[[.B"k`5Zq:qKIWfܨp@8 82$)Fo |ʜ#[* a bg4|k9lX`H 0Ԩk9#wry%cj#1 }]X_&.X sh2awX+aco2d\v:֫G|>i nz͗`4 ?e1Ư=zKq.u.:it,.:z86iIFSMU8ʑmsy/?.xsVK0b؟>qqs@h<,cj&U'*JUBEcxeۭViHLPi4HԥÊx8wXgS-S)E s5>Թ gx!q@ :۵gw4SwsMPT' ?rw#wH?I8x#ϛ6qnΫ)mVNGY9v=d1ղa(x4Ӵ[U\H v}6G?!(H{ĝ|~}%1Jy܆xcXA\2*VpY`k"B,S+)eōs*pHctwqmTt|t ,&89A8"}ҹݱ8y, PA:[YETtAN6 \8*ԆHkǡ#Ppޥ8]Lv4){l > 0!XbG<| ҰO^xE^4.Mv\hF^ _>VU\x6C'a()l>&wG;*E6ð' 0h2 Gy`{"S K\S\xmiJp۵7b(F oe0/7?e_Bn ?a\ʫ$ ɧ٧6v}3%I%IXJtR3ipHk9Q% ļn$НE:cX(f͜?g3㻛}{?4+!vN(}[}ےGztIL|:rkɧ& k;㵵<\SUO2IJ:V FInsNh<7E^ڏ']`dsB *~Y1sZ/FDI4Qֹ=rb޷ImnoM.ɑL!|H]ץRɖTp[⾆5$ukvTu&ʽ{$&F_;[[.q*lmy藠ε-9}qCx0pD7QR%S t&N  $]%^_vhO/~), |KFr[5.*q ZΚL KprD,1Wxhyhv\p C 0ZrrxjC3dih,0Nmojֿ09Ǜ yΑz`0{Q , }ɏ)a,A 9F:3ӭUUP(gTHzWv!8ieA,WT!b/9s"Q*KeT"m'\6>KpXCl%dZggn qeH0bS&NJޱn1FQT;A\r7ŻQx`we1UOM+m%cjZh=M=F<8I38!LHt*ySqKx1+sqIr=#cdVS&pI_=BXߎ\Ze%1|gS}{YZ" Jpov_Ͻsp-F^>#}/?bm$K/mnՒԊ*N%}Jzz"]6/3D!یJ]6TJMEo/pݷlZ5C0U7%jҗorEErx~ڹ>p9,[1% RJ ՗M|v.\ !-s>tk{GOC:<58Dt\%BL~Gsv=бIJ[34>Qk1րW V1IRe{r~Ϫ%f@zImN%'uX~zg˽C$2>|YrE4 $,El*ΈQd(+m6Xs:ͭa\Q2 ]y͗<&G ?]59ϰY2 v4q=l2wqca-ZSE\O#0`fp355~4k:MAZz+GV,Pf837W,(Ԍ|Ktܽ~JJ+*3Si"<-)3)]5Frj9wyat0e`kDRNP(QPZ6Y2HI1%)HǛŚzvԗj)DQ $CBi ̈́MjHxA7TjPjE&Q[]?_O+A6uQI/գb'EǷMQS#=gN)+PxR &H\fQ<=۬~:wz s)G9LOUiժϣww#e%~\KR_k2tȮveON,325 @dQB쁄}0;#5zgQ3!6k NʞnoÑ:̅hyy 6dEsF/ڃLYKjv|9eSIӗ[euA==GV4cYNCay5rŤRҴC?I=rо_~"-@l-3H%"_^ܥ6? o& 6L x+٣exBIZ]9W?l![zY.Tr\<ݷY熛<$1Ɓ1ATéۚ{@l= 0zI*E *8i'lR@.8tySXuRVrt(9e6jNF =r,9RZAkDm d-=&}Kǘ<ɒ[<ңIXGA0GWBBUL .o yun]h;7EܕAV *w<$:ޙoVRkj4m}`4ЌW` D_$2{hKaͪhԞScT;V^*4Xlb- Btrx5n|lhY{ذE;auϏNh9Yk$3e,ֺTRu#޿יwUp^XƄ24""k}C`j,PI3]+ A}+o㨡=7;XΘ:WLyPJVb:7m9ۉcz<֢eSIqruWq1⦘ XFM5Irb ^}[!=ޖ5Z)׫LD؅5՗wK%K?I.8JAX00X۴Km<6ር?g-:=*W?Z<W=y2S&8 otebVp$ [%掼ɝ.'P^3ٚzNP4(W-GK&yîi`Xo]Vu8FfiUg@pLUx ʙ Rox ݻn}7 9[tP (\TQR&Z($z.Z7ul;@33bz֒-czNbsMݳmz{1F'`PK׃Z|_ȐR#:`4d9G9;lAjB۰+ZxȘa,A{HقK2דj {~ (C)qWo^_Q NI@g||*ΚEʺy3mQ.[} fĕJp 1ޅL~=_r _Z^_m/}#x%~-@C_*.4F]q3l\{|kȎj60 }ZfSl[ ٛ*VtQ‹]&8w8euH_H^b/)5ސzRxv |$G}ro9z )G}񢬚B7EB me_^żԠm]^Z0/P5m%9bX7;KRAeCe|*v&+L9T,X/`*iidZCnIRmqGTШu^jΐ7qM(/,3I)!%gӹ*W\rV4ΕZoJ7:EB,s!x׿fcP#kqZhq=jn)zBL!Y_x9/:*1n-~ }n̟Vcr&l$p1VZ"/奆fg _r [o Ư `-G{l!:]À}˓}(oaWwZU,& [/~=N@+ؚ=8}2b峌l_ח׏X_fg]SEfLw\pX_Oqp Pxz-;@,F(dF#ҕZŕ1YS!Y5O#I\N( ['AUw61ml(>OVdGcH0`E%EVw<&e,1GH9d1E2J.Fed@N̔e)3)DtG50JIDkS/6#5\l+-_Pe, _~t.݄ua8ݍ/lR(Wd/n>ݩ~(([6˛}3OY$0iOQtݹ.{kSB{wp3*X֎Tψ. :Q؜8m <8GϷ77Ws!'+^CҜNGdyF7ivK#n2s4*72}c/$=wDL!9rxvI>4JpdQG*M$ (49M_0O-MoH>޼||"e?}r"i_Yuch !a>}u6&9RvJtԻlhmTxq9z&bs+[b&EI,Y,$!u2 "!5RcQ< )q`c "g!<,4-xR/G#9]9it?5kR, SQOE?TT:@F(Vmn}!Y2d)ʐ(CԲ & Hm7o솰sYybV}{o/[˒$‡}{r:7%eK3*y|h6+}lߜk}3S^1vw7n:Qr\a }\sk ei_.ZF1{`yCk>5(3 X##0#AC'p@#gk9gk\=i%2Ώ*gը3PWDZG!|㘿 Mgp;ղZd.D2˰r\GXi;jf7{4"%L/Qv`ngfq(Z0Cfz_duf;Ϡrjߎ[*fQmn5wsrəQg\.}Au?HZ6@kZI[gI=&G&or678&n|9[/ϫyDA*èB(VLd=SIEĦ7j{ D5)UUc"! ЀJ<3=;TJhԉ3YK8oDhP ȹ=bT@?|ZW3HP e L`V*4Z 2FN%CN$id!VlGFd'i,R_HD-?Zq3ˆkrb("g$Y{)ty.bx(Z)`Iu:͍Ak"-. \Fxa&XjW+$90ٻH+~Y+{y3v˖ZIm~%YYV Ipee2 2<3v+ R}Rf yec,0?"O%_ؗdxރe0XTE` Zz#$EȾ6DŦ$;iRcjx>F8zU4qZ?[fؠ{ fuk>%GQ.9c CpQ`p^C2%HE.dTyTHE錊sI2^y(ls܍] ,x|]IWgD™mOW|"<=zA:=߿Ikq&=z7_kQ Y, 8ڒ} 7n#?3_^a^nI`4 Y,`{N(~K`/HZ26o-o+ A~zjv U jDτVE:{.6}|}=Odi 3ݧ|8!o1GKM|[EKoWnf$iq(DNNg @RW^IBcF%WzRk|c^H&ZAq1IRz֓4"#d2zO\dґ*%Ul2:yx竻{s>jcG7bp{cNjj}>ѝOOɛfĜ.e"փPb̓hb=J '!YbTڎ,J<3TgMޒ ɇ(`l 2Zg<+QEeTg^qQG66îaZ` XXȪv;]arieXCX *ZsH*7:({|~;Wg k'pݔ}ff'cƭGRNGt"3r_oCi߆hY= x܆hGo{U![) Al%i6<+-AvHȕA~q~_ٶptYo-L/.q$~9W.| ~lڱJt& ^b|r 913SVf *lUHݱo5_sec*!uNۇ[zv{Ш<; RǪ!=Bu:+$&B; :!LEtGYRoJ952,ijrRF`ߜMQ$A{MKK׶oqcl@cMr0 턣:,><5_\RKNo4[(iJQpN>8FV϶dgK_$5z Z%hu?zwqU[@ԨDm&:ʼh4n_77nyɝ$kĐ9d&e?2=ӌeCiTVcʎnۛTqcUփ  fY.Hq^.~ixj퇱7sm{*#|ǹU:\^H&6ѯ;MJ#USPY0RtT#ߊk*X% zJaV%)!DcC >HѠ<<ˮ$=v{H>ߍ*pBOw_P ;P)~E 8)~z)aG~]X-^gFۃ`39iM}}Rz.~MC6^2fhAZ)15ݼ;8_7h34}; Oi zʽǾƐ0Ssxv^J? zQ[x%O)B*JEZ*e$"&R6礀h1/ʬ"3^с&cM<紀bqR3EZ%=],دO%9b u pX6Ε7K N>:q VXE`z!:xIYMeBg #@`JI 6&*BTBU4 E̫U V:$PHUGƊb'srw J9/( kɎ^*R(**ᒄT9g;䔜 c)G 52 \Nƣ}Or-A^xH19i@v=3tP;4AoȰԜz!u\lB\Ԡ :fy6S 9RڮR3aouO஻\Zٞg\IS:`{yjyxX֪C 3MMd+ bVxyTIғ`|+x+]"kMueB]`clV+=?Qu,{uw8ȳ9ǜv&۹{Gƛ/>d/lu: EherYo4d %}%9ex(tY |p5VgQlRع4=HHN0MΙ3ބi9s~g֚3@V&qSbL>u:l;61k⋥&/fS˃־`YCNCfoA'Rhq:B5:& Tu:I? xX;`2U5)y+bon\{Uf6X n-0VU1)Pf嵲H>#bfbFQLE6! -dJ1 5Mr%1 HsCKGö8);^ 覄% )̬ giDc۫(.F\(uK@QT".I ȟڼ<jM)ЀS5*/œ[yH $x"~ .KkmH Ogۅ ;ɗK S"$%' ~CCq3CR1 aO^UդShvofz"o9N81 wWbl1 +B(t%CH8 b 8Q Ēc*, : #Xa>`bSN fD>GxI${@WGbt0#)9} I٭\7_U*I0Q\uH [KV3H1 BU"Lqת ŋ ^Q{c,t%19Iic\N#uTY1Xmr8i P1L]Q Ye iP1bwuY #"[xA[ᬏ a jl l&#P4FL9x tWL"0'H2c 1>e=duut6)qK-ZCi<%MڀJMuؓ.zitYHń*N=C U⒜v_nl]R ׭@SDgK̭zbYZꗆ昜LH4}'N}_XaGL0;3ss7s[41^}2TN$[BA~l>{рFˣtnqS=wym->r8X;juRg$1TS֧a)a8{{`_Qz5F}e1go7=?0Mz[96#pV՛F?ÑVX;efҋiO@S#ュ܍'&+_iɞ3 { N:͗Cɢ;,oZvퟴ}BЫQ0Bo0"z%F#+ͫd:&F`8Wˮz}<eJDF5PQ?;!j bTuUF =Bܙ(_ k8t'+lJ8'ceC[:x"LAym@F*)lLKJV 㩵 S+ 戾6T~d5j w[F@۪H-,HLvۨ]W"+ bMl6sI>(:B{m?Ds&ZCCt;,ЛwJ}=ߩ׀L}e "1&58 .H7Y=\z:bu69N3@DABt.A$b̽q(/]jʂi*V89UEruεRm`ax K0VêCY- PAN3%bk]TE>ǟO_(xp1K07K!EP]*F\"PVb8 VQ3n3ΤkEB@5G2/ о*٧/w5vEY<}*[lL/܀Le "s) v0p-%s"kdhEt{32 g0gހY%,MՅj}2}O_xRvt$oB/Qj5oXkp|ۄQܐǝA Ce μ0{RMLUD=[[=U4TUuB`4ӑ6g;Ni:ݿV\!8޾9AjcD_)Q,li|T4B^u=^}8þ!{tKس %{o`_kCZ,Xu%T{ȡ`_⒝KP &l. W{uujˋKoxK"V35"Axdk ci۾%^(^ɗݟ4ֵ wOh>˜hu$ܪYxgTn ä7N'ڗ 53SL{|uKtgM&OvG䫌|0Ǹ~uBq6ܯ7@s::hwh}f= ʭ^D$6`qNCsV: R (AI@0 V;]ZT7:6^9NӯaPN!a:8i?f_ l4_"Lf b5_}zfvc~5h֛D._Bp JrPU?UL/*CQQԨ.dқ1L}wT`X"5 >bHXEO1TȐAH8YU.F"P%gЧQmI|NL~lhJӨlQDf>jI4j „6Qx#."c3–(4(H hCJ3[モZSsIƃL;hcɄv(Tw~ee=W!T*,ʛL$ ӂVЂVfX`Q0Hn<l.B@L=hyUTiwJm ]I lP4L*8o$D캅b%hKfe7$ w J'5K,|*sBqk2C( X!ɟF'~3kp42HXhsMpT@n!#R!t9N"(|bynY r˂OE%DD8Xւ;؃O"jS .iFI^\y.J-C*fqυV-ǨMhTXw+"ʏ ] !LbȌYB2Fʔ"1B\ʽx2 UTzѕѣfÁͶs&YJz2P%U}G<U=7Ľ#0"N%{= Fq7~mL}jYrLѩW#[-*]J dݯL m̊LkrPMҕ^K\|nW{RUrq0krFG.~^P:쿸AYt=^sέ;vDBN"k3$E'6eĥ߳͠N q/^S)\;%q *2@rDǥcZ)esAy4jܕh-LCkBau?\ugTE%_9ѱwY@nA\N$1=VG1DZȒa 2">)CAع[N9 ;)$VǤ5TӑJ_3:)/Kgrjr#8ÿ̨ s1P>%2}N qzwѯ5-]]w@_f[}{fu<&{?VL"8#S)`R 푺=kZꗀ0K/с즏%ֳ h*ǎ y)*{|؎ F/k"DGT(-J6X>]:EwY*:Jj` EM_j8/h r:L:$!sr=`(PtzԥAakiP@d* v.';콁Z|'$w}w\=#텣Ti3mw =`pִ ;y*@Q~6o11(aQfElq{gr#l)^~^zU^xkݨݎdVeo>^iE)y5^hI_ ЛdqMQf^:^a|^8)tsg'kԊ{PJCUjL Q#nߣ{cII 7Q4|~ͧc IM75BRjF_hBMXaH-!ssNڸwڶt1 ĝ^ 7̎q&sxG~kyst< u2$4*LԼ3s tYv\E,y> SO M: cL ,RsJ N΀{PC.23C_8,M="Ux(u"e\xUJ  V\kx5L[ yb.H| XT /y?JJ{gz'K}Sx5{˼fpø 9a>9ί\ xBm)!`hnsN]yihŅe*GB:>/ `)qYg[fbH=!6cځLF9м  v ׎s^InuRZǎaQT(1U73Xʬc J#+RKΧ+#]4P]>4& 4iF)6#"yԋ&9-]NqDxj8Eu&sf(ʎ9fSV ( ASҒ`R$ UT$44X n}BKqs7*ϕW":1L ǫJSϛ \! D]NֆM1pE77@aA`<^nuZ/FO vM4Yeӂl-3|mzU[ITxB;|'wonhe5-D@)z^bVXʀd2NtVqX5e^Kf#$p颌"JyD/&<WXd_{Mu QY,y#{Fc #&qC$rEjn&E e!mo%}8XdDN CרmAh@{SbU)QE5 E&O3bnbV+Y{=p'fԣgT8@0t : $3`uBs 1KeB J'=Aԍ {,^H==c^FI?eۯg?ߥG \V|ǥ)xn1iQS|߻yM蹾{V#vk "x{x`Wxq}a"ga~,].2|}Wi7#wcbewkdpsgȖaWk?|{vOP tA\>{>F8h6 ¢iǭfLnYG2Rjk6j]'ޱe45 /Zѐf{ANhR8Y2W_8YZ:AMU2o |,;& փA~5tt3^Z+Cx)ztnp~fǸ͊:"I"-knV^N%$15G-IcvbEn<|?\ŧz``+1-Bˋv͇6o}nGڍԣUù;Mȉ].9l{NЇ4R~W{tui)IIxPGW{I*א"(v'NEs}I=ZU<$˟Yj3"T'1X411OvaS=~ŁAWq?(,GiWW28'|LIsfE dZ꡴qL SXBAqѧpW|0<@_CLX!&{lX-L\UI9t 6CSij,h7ɀ!2 -b8Aѡ 2Fm'J8L^J}3MZ6^j]O^_@@L{J#D2r1aW"?`WZR`dʠE BĻms6q0dP/=A':`w]9Y͒023g/%n}elg=&`@yo?  *b1pv_ޮ\;V}nbwSQKkeO} W倿:I?,rRm1zY̊crxWgՕr()Tz!.کO)=xǻi![W JNw4nNk#ޭdлC]KӞN5nFzthN90hޭsCK[=E)ȅ\3$H^kbI>@` ʕHh$.qvw *RgHܥ8hY$1J*N!TZ W ҚFg6SLdZF1yPV艷(8Xr1Rk j㈩te׊n"uEK7vL<1$vsψ TΩ1ւ!)E3I@K .xԢӌ+Rۑvc<_bv ym]?f(?~ſ.K˾b_OƋ1߱ϿO):4Ac'^{Ҍl㓭[|᢭>gnH(9 wJaCN:n;HnT}Ǩn]Jp[@Ҟ}ͦi 粺۝ r99! ̐O3K" =TYd!P %n<0U"pmR"{DA%/Za_&$PJJ!pNZ&&U".x-s>[`MHĈx%ARveuDKϩ)Y4ODE#MZb_X9%nUJ.LZN#1>ҹ6B"Ԏܝq5vщxy5{eC6D@r]yŽU4L :I2j9j pT;]ǘg->y=)N J(G!)Ì/ O@C,FE$py멳!jp hYLyȊZ6ԑT"@|a,3.Qř(Q"SD%r/!}*-/*9!Y hm&(&ޕqcۿ",js_;d8|qШ->ZKU*b%FnX\]`N@)l ,LIM(q&gՇǒֵ1yueA$$eRShz߾gx6|=*߾D\ D%qhqX vvh\ >ęF puU!&*:02FyPPNUνJФss x[2wb:Yב|MJ !&TŤ"߯C2вzEטb ;4 k 0+ g|.]:!Jw㢥ЃG0}NDeZ6,d AB7!hQvD垢沾|#`SX?lf:_Ut<|?_Fכ'?!!^NJ /r"Gt1jdڽ|BNf,Iͣr2t}7OK&/[ڋ/Jp{lvs`x=p%` @sA@ xA[Ϝ@@t/Xo~Jt9PWTFjf*ʒDg%xXF,!0DE*"Zl,T>a(kH34ձ֒(a-egeŠ$MWYu~ZS;3v˂7B_ϏNNջ(iȫr=@_пF_#r|=)KA q=)GkjN$"NydZwf~r~T@kFj BL=_wŧaJ~˒ ~-_?G5J>Χ"FMRr~e%{ݏom^5ty_׹ .f7zj~ADDGR??`:.C7; ·{Owa (,Ƴc1j 3h @sXz :v=hH-ԍKr@{_ywkx5#X*v ۼ;O1 Xe6\Śٸn8!4Cku1bLY1bY/(!D^0=+hXۅS04T!hoz="ԩ~6; 9Ѽ1UBPRBt(iX+wS8W8oe4U4=&~VTKMW5Ď>-n-WTRzc%҈J16QO`Kgla!eӭxXݠR3Z14ѣ?o_"iZ"qs4H繛Ϲ9b\K11/K2qT2'&K( ʸ<=76\/ ;@- *^to\طCLp F8f^_4Br`h=N$gmg-|OO;gO|CtV&{άF (f5]PBT fgj.5Ѫ봣FG[вYLDbR+.&u!AԚ"٭ͽ{|l9پ푡B"ؔ9@sˆՀIwaOj]H •+>eU3UU$uOݝ`|wR9/0wW}(~,Cllf̿yg%Nc#ckJ~7F|1użs"s>^o#i?yeٺ4W:`;yؓL.nN;n^:N ,Mn=g:4WbUg{U3nN;n;Nur-܍svCCq ҩBnNnN;n;tJ9ա!߸T!Å8AcΑv] RS~֑ )..1){:G.'1E }u<ƂîyzÏ3u=ũK*Ԩ(OxĢ#hgb}ߔL2cLdKYfF{7?s)g@+3]PˁVBz' ΞN'.~/l68zr/FUw.ACT \H3ځ2xXà܃j_D?^V!aRbISU'zTUםk!LhG/*A>r_pZ&vqYIo<|d_=sK4NpZjꝧq-j)Ui}4)bK *0Ha)T#lk"i2ؾX( gT0;U(߫cG⢋ J1rXd:1 K1 R31ɔqǵGyQcE*L BqK/ˎXqTzc,6e}-&{>;aofAj6?F ƂΏkYߏ~tc^q jan(V&$4\*E'2I Ib觎u]-̪Drqbjz/rU#Uٍd/J*^(dǗ'(ĕ>Ue}֭O }(lk/_im O'^xyG! *pq.JpEoiw=*XQO AiH qe*dWedv22^+#1.z)ADҦc;['/1AfF>}!]S O%y@Z:PkZumRL: Uut)vq>[+q50DڋS.Y85Tnpshk}VHra&-WGP<&B~0z!]&I;o{i}X_Ӿz ,hL(z5.x JDPqEP]ݱ f\qY:&rY"@?޻;.)' Y x8hcLbDE*"Zl, *qLkj"aTZKD5q%S4U ZX[ 2 7PWjl$r^}QYI3T@;! c8e\gg;PyjAhɣ/f ²T*1IBS& 9!^*E։&3dr"4M \~J YԲp|>`QL.bү)έoWm@z-F{I·߬[cTSL6?,yiσ'4DDp,}_a}b8(/zƳu;V?-Fn<$unydw|#%H2LK7ލnp<ШԘs VV/$+\_[U)Sٓa)iEZ)uͺ4.mF0L PFU"Kw=76 T~FSɉe!RN`[I\aXbcɘ :#SK,{@ie*!TeZf|V^3w X[)ir?M4Ƭ{h)~xQŸBo}J kL7j>Ȯ[49֒1>wSJxknj4OYbː39I"ֹ2)DƳ F׿/4)4F.ri0Sg//4s2Š++ԝV~7 [Tݢ.wQD4y&OR~dJ l%{s~q-.%rӢc2衅fX ꆻwu9`S$g4D:Je ,%ڊ]?G~qt0Brsܗ'?'gj:f2PFx]^͜ϕsa+wIZ׺v?Vϼa3F#y(@KRL!Һ[@ m :S?_J}cjM!d/8KaA%1'Rd*{j[2 nSbT (* w>xמEE>}OGl;(w!HuσDJ# āCC(-bFgY4SLye͉$*n3y"Ϝ?iG^ƽ${ؕӂb:Y3}~Owk/ARԴ9gByVw(z;ch6|=bg>IP6PB/ iD;EJ!aJ{F_a30yq2 Ѵ%Ex[MuYxD鮮\원BL%H>fcFGR&T"Ut ҘEgTs~w'*;vw&R SWgo`";C1/N!tx4Åf޾5=hSFm&uu|<9oYG5فgӻ3J 1=v !Yf0HF:ȠV28QL}Y鐀@3ŇgFqn@(W {_MI\~_Vۭ7s7S&jG=GN~N.u8?MEywW?I%.~?7;6=ȎxvTGU;BG:BruCCk|L. <K}QF5'h罨{=۫1'G]ϯ>|;3EģkNpcq{G;ӛHKy U2-?}zgoh4m[-&^~u;Kw[NHW2_g?7IS;_ۍir;ꜟu3C[G2qѽތmMh&76@㏣{3#EoMmf^fɟ|H\p7p4gFKtO 4 ٨҆+S)_z~w}% 4brdDL+P@?OEkD'fvrujW.B?{3~@{l4hd~7߄AT0VmT0x\?.|puQZo_0}:Z/z݃5qkWo5^\A8k.~G&U/ôWvSL;Mx~@OHJO qV"8F_/waڟ7#f9W`LO0F$?kDGz^]J^f;GAOo LK*2 IVKf D BA'2!ӴT m̳3,5 Bg˂Z~ nLhfXV% 6ꃇ2I|Őy.<ZśvP;?l7kh4 WyIJ\/7SheyNSa.(0Xq*8}ee:@'lh9d 5\3m4:"5*7`hML&1o4~riÔBSk{/՛J2^1i7 9NR\C[,;[tuF7duR-7l\3[v[KTc%UӳkgHU.CD4#n_e|lKGVpTwNj1#uJR,\ ڽrsf=>L<O3ELL*9h?yi"99!^7i*tⷛSŬv4d ܏nΒ 5e4rJΒ ub*)!MZ슕&Bo^)k^)߶S+QƃH1iIO`Z=&EZi bl%A4E69_9^tnZܬګs>eb*"/6L}-EU6Ta T86htWϥ;d+4aΕ:-eh[yͽ];Ũa7[j6>v(j,S ZqlHe7Sk oy@)1Vʊgx/­u!͘k4-{u$Fm8fuU/RQ7)ph12|혵tcKQWfkakyR-8l&ufiWi58kܧBpcw@cd\.tJod+t[L&Hg:=Yu7}/oD]{j.$ Yi.=^ q!! "_!Ax >Uf,& l] ̈́8[6z3uvrJ"ٚ%r{oSKDzrBbv4Ë/%E'0f7aMe]\vr!@d):ͲR]fÍq. %K϶ BQ*̛dݨSddfLvj (+h`=⽂fVWPqA1Q+ёxv"*-(i#` k#^v)J=[ȎBwlgmo#2ɴoA/3@q}ڴ=n$nnϰ Lޓug'8Q%b gAI_|܃M'W ]msF+,}BWCn]ݦlE"$ewz@_@"-:Ȓ@`tO3=e_>lDC-zB VVA<4ኙwiNJc:)ǚ"|9tL?%g| x:8"KH 띱fǓPCYb(uβx t`t`DW>~%HC[WKHHOsw&I,hK*yNJ} 9XR\ )uĨ$1 W {MZUjAJJ[\g`$xf^AZȃ;% ^:vQGO "@ S!n\{lY`ZWdtEA&lI,g%{ˍx;{2urs"N208qp}:"[$WH9ɗ\|#0'jJDjfg'M|Nq&sV%~Odf|=ָR%1\❦b`r4y0̹d=g^L'2|~+ zD,1ΑJ[:"f:96Bqb0FatFi G14 Jk! H6x.Ai1SZLҰrZ/r$ SY#&JdD"Č^3l2"'Vry,6i!ꥋzUS 9Yhcrb,MvqiHT88E$`II8S{G K0<(j$ł;cG#f )4(QN@\ ӫxk n(X@XCQvd S $h- h"KLl f]@7)1lpȰB3g8^*x0)|jO: cQY{v|/1 67I Tc߻rMbj{uNuxR]`l+vօfĨ9첌|#%uV:E=E,r%X:fq*]nwݫcb0i{!xj: ;ǹP(PI%M4(՚0 @!T Q;e?;3x7qrA< 5- (=ROlznuG(;Xc+W4-?ԜkBi撰Ɩ 䖛+a4zLmXRSu'RlZ>t `FIq{˔'3yq;/~^=ųZ<u9wh򠟈Wc_Vji _vws?-۹!Xx{$ ʝ4Ѕ巀t"Yo q&{ݤ}uSZMQ #7l JI &!\d#OSQ9Bhd5eN/rv2#tA/t/7$)z(f3з~zUR<S>5!/ĸ0hÐYdNXDR@ FB`B#%#uT3Z>! 5elv"OaVsS8$0J鿇%\Efhzc[ uax; dBiWzɀ(/_ Dq#*pBM- ac|. 7CZx~b?0F{Nfp|Ytݎ,0W#%z0"QQ=THGb@A)(%`SY ;P$NHP rbFL&UgZʂ%hvZr̙x6CW;mCTi!PBȃ!Kރ 0(EXS@@V<K__]=> l0}uǨp+5$kRo_\'Dl.k!^C=Q:\4xNdڴSg1m fD+ Q]3uV ^ePx#k$B`^H.Ne6h:7Q2*)l3& ,|5PKxݔ3J&0wgbcB[+ m5&2h%FMJcOJKE:@4VbX~2δpK{]!!S .qRSlNhG$I<%r"15.oP;TyO'*xgDɍqj1 ?UK >™oPt, V؍ Fʩ LgEtax4n_+ߣ8-)K\KmO$g?f@,RqQqMQlP=tjї:9 s-^DDK~39 4`y(F²-Iņ[̉ehBL[AV\뮲O97 ⧊.5s7ﻘ<0=м޴03~(:XYOV'V&L,F=s::!I8b N k! 46zA tʇ#@J1 Ng !YqM逎\#TkhRSJ=ehQ-cОG#l-讷9K;+ܠR)@z fՑ»' 9ǽAf.x~yL|B/땛4Zh>RռnM$7D?Iq;܇w' 2OmQ͂v)b`'ULa?PBx([x)T>%[v-h"N.%%*KTcvRnjǞs PVVk.wɬJNʝ֗w8KmȓO`.yDE4üYs`4Tن`wy(i S 1y^9ZuX.&OʽDCߕ^K+m[nkY>Rm^ochD eeJ-0ލ*QϹnZ UEۛ ɑ&.)̪2Ӎ2tY՝5N.kdes+{{S>/!Ɵr]m7jvvyrZ!\m,[VI57ٶzf' 䌋ɮ+6Akɗ( X}L}0{oJN ȔuC DnWql=aݹsHp`OZF@adiǓd^8 ~^z :5LwWz3(DoWeM!NZ6__Ocүyc-8 h t/G-{ƥBfiRk=GPm,臫wrcd$F-NGɪ 뼹N_^lۿ_ G^~=@d.2߼wdoGI%?T0 m,{ ?Yq[_ K.D],OU ]K&\y 4NFwe{VmxKUN-r*OUW/}\`^-YZ@P:ZLa|?> NwM1^ra2'jTyyNW}zc'jJey\zCy&eWӡ;5j']ӡNie=ljn=+%>lrnEι{J DLP(H"6s&AŶ?؄r"I>m=BHr;άU31+D )H%'"g37rºPm薱><"fU~^Z+T yI}p]>Ֆ)9{R=[jv>;Kh#mhϛ[ȃNWMIL' 2n"ba%+@Y` Q1!(ÅSBNQGSWσ6<=IM'+ M@9#!,< BLg]0+O N;ntAm,BjlURs:O L9ěcەyЄ'Zxt gx8p!̷͖6P_p%hTI^P,H1u&NAy4 Rnw)*9]x5Wx>G=穑o 1L+Ik"䩞eD,;A@Ȉ1(I 3ge0JNЂg.xƠW*Fޅi<cS]XJ4 b DsG"-0~@QFIl|#Щ4] .ibfQ#_h%U6*bLPyo 4N'x\;*tڭh:&E'k%Wh4odo&)t<ʑL0DPh [ +\q\;Bi.MGw(L?Ќw<~4]Ѽ˷ՃK x ˇ<'x=9(AwyC!AC}h^]sz ˛Iə?u;g3zܣy !JYdnшK j,Yd3lN@vOn|tMYS_MngC1nzFs^BȺ)z dVt!6 Zg}˧h\,j$.؏aju{ȆFxgȭ !Cf];=d= Ys)zsOaæl: 76yͩf]uԯJyt⏰Xu$~XAS_o?QX,^\DVby.w)m3R6ld=u|N&du|㈑D$6,?n?B%0Vޢ=܏f *48c~PzkՐR4#VIa g'O߭":\;R00 oo.rāR{5Ν{Yl>y4h$`a:Ye@J%QT3D-cNBa`(jG댇O|SƕRgo%,VUa_8:f#IZpJݿ/Ų)8>,3ԥig-|cdainQW"v"01aFkje~~j_^ngÍ:z#>_nE"s)O(w/WtA z~ ?OҺZ<&+U^Ri[ť]L- 'u&E "r預\Jyû6fi~JxTPTZ[(/`Ǹ*l3Jt#K6K.t^u@tI˳.Yzs]GP]J-Еz%K1#HCH6)."ٔ EcqWoe8 HZ~+ >`PިRuƣckJR606 );nX7j\9&Ζ+QHeJAf8_ɻeQ\\]\ʈ?|E$+MUR_ ;⏰(ba#nE=p!y9wwKifv\7?L{䢉JӔ]PJ1%~ i6`W3Ϸ̬{ب5 T8.COLS9vc.QxQ~-yZe.-^i4/.~yO +tvu} VTS~l䈜 8Nmῄw ~߿]]rtytep+^#~q8N [?#01 ]%*XPcF8>fV)dMF+|%#{A#QKW:SX&4D*&&i!&rf]eMjVU?*<*ѱzZH[eVkdtAi X Clc PyT%0v&>i1w1&edXf&\ Xt2DEB- xbvEME8QDt4 ׬,#xQHxU8%b4z|jg$.7k:1uuQX ͙fDq%CU94xp81fK>0:W_kުrzlRҞ]NװSs [_kf(%*X'RgMfbh@o%T$jH35ITCcB\yCYft$?k!Z`bԃ^D*QHb*~mtIEiN^S焄X,*Nk Px%ӂ[C8ts0Jv (AUh'R#aGdK酡##*D+ R({T' r##Sd$M7(dGF#)❂d-L׋U77ȎaPKRCtm ,6JIU?ݽy0Wi=[7.çliqBulڍi܍ >ݘZqZ:ݘZR VJd%uc~%uiOqFmbponZKlLg-2rK*%{KS9Yv%umSa}Q .'p%V櫵EcI*QQs=Q82.> Q1p%orCZ'?$@^!lO=̽q&\`OOG(,pZ՞~Le9}-wUl6ŏ9*0.lJGd.ar`?lq!b+ ze?p*1DL%<2vw=A"ͻ3O+ \jYxxƠ!WfXY7{9ktl|8S8Ϯ>uO;`hb_eM#}WW_MZr7w茮1.@hk?LFϟ>t|Lj/6ZSᕍ2|hlm|'[WãBCl/$tHZ ]uP^Nʪj1sJDk=uvϾ+["<#ABPu%TBGp_*|᭽!>Y i B.^ƣ?j)8Z7WnePu ^4NEC=\s7[)UKhv B! RCX+ ]N9g=IE C5K Jj&B5|Dj]J-KW Á_D@Fn*-9.!8 @j\ rʛy^/[ ma:3"I&JsZV_^ͯ)RrH:+{ 8q`#6B0 Cm,>A#\Cp0̟_ OeUզ 67լhG`cQQRY_j#^1ॅTȐH_NXFOm0ԑj4 Zo4#F6D% ?( ;(PJ?.Rk t܂\8[7X+)d_ߥ/_2 ʥ-t?>DR бoqF-` hLG R-_#2a޵6r+"%Xmbh`9ŞE3g7O=ė9 }IIۖZb_պx<_b+mr-92dQ&8p2Z,k"i'h piB;?q@ZʥyU,茍}3GaHsG?CwQZcP$SD(7gu BѢOZ ~\(7@ xxckNE_|Pn'OjzB{z[ώ1#M_|)CJn۟W]Jw.$]28Q==İ&ħqΨ)5Y <+ׁz=6ϠJX_?'KbzޛzaA:o~jCh0/Ɯ?Ҡ zizfRqɘ$)I[ "i⥹&xr!_b[g|<<~{XV^ dӻv^ 3,*j5EWDڰ}B#aJqtT3]Ptu KRZI7٨ VEx>ֽ SƯg2\rja IJ<25sdVƬ5.QJ Nv`׿֝*;lԃ~{0}^wDA8]dNi(Dri,T.KC2KZ" CRhl4bЊSa_h%G8„.iEYF[cP}|Ҩ#!r#p$Q]Lj0ﺸ}F\1h!L?r3ixWKQ0Oks3i]~zpn#::1͹NSdOuuT\I9.6"Riі?MNOV ө= jq@x "jNeynYB< ɀHZB[dah+SN\R`@tjА* 29ȋ1`94#F3Bdv7m|u7yɧ7eDr2%7{kv^~)&;(ZD6^HYSTV5o7zιhIy5%Dz5iS*fO !FI9jLze^/~uY1x*˚\>e8&4߬QDK%*cZQx{R9a=-!D*tf}4DibSH TVM(&`4T [xANhku||aF*Ki)8hBκM5QDF:]e߻/ޕ^|'*c4~&ŵ9 p'D*Dܥ\,dҊLJ)32ŢР*^(ˀI;ejV-QJѤ)Df/H Am,R(<%d+S0377o: + ue,eJ4E`zk+Z2ME[1U`&jH pڟ-6zᇳ4tU*Qol_- Z"<κYQ]4cD8Z#\3-R yL88 fFrz=[Td,ŕ쑷e_tQٞ༳C-G-Ƞ51BuodG¤հ14>MƇ`xVR#(7Ne+c'U;/Cocy~$;%9i+2>Ao2T&;a?$G|>o#~z5GtuXNT%g )/ކuᵶ& >?14g-ze笈z 04afKo?NA^1PGCDr#FCRЖiklVp@)U]7*s@ۡTI2tZg52 ;7Nk,a0:uͺSK JPU+Vտw00(kS޿I1Y*`\zu%+\5J/L (T`l(kU agҴzuu)TJ1€BtSmKt[6yfr֙Ih$2gnΌV|Owl-)kIAAw|_]el'u=#Ӕ{r̄R1I ט@ǔZҼ"RN]72.n[j!.ـMopV,1ZA}oI-Z hnGl~-1}HC,(7%QpP #2 M,\=̘{/> % 1ƭh!M=E$kr^1Q06.vkA>(B-غ13A]Q+cvWlrېݠzu 넧>jbG!:x!Jt{nvm[x`Z2Cz3|3*Gf(_b,:-ڢcJ)<>QXTq+(7#[p-HVFɻ= Cˑ t?s] !~Jְ׬?N7J٣<{7?Gw* >:J# "zqɘ{<^=Mrr4GINh4yǷI)ys i\=9Y @]kpvWΘ<ǯ=K;ۇj<&ձJz{0͈㝔(TqB38J:&~%g[ IQR!½h؀.XK({; P,zg Z+A_'PG\5;!/y4ӈ9RCZmD:`+8y S)22r>mq?@U6%aH(?0^tlVC45`̋wo9Njadb?WOyiոKh2&)v _n{֛l oE$'7Myrzܔ{=N?c4eNSe Hs~!aK3BlfJZp0n󢰓|X|^Ls/s;Ҥ@דxs_['wŘ'Wti2[؎ bB`1{eq?#Vrpk7hY=H|UQ2͵}*,Z_i`̯Tbb*$O O4E$&<1"u\:EW`!m׵y וwH%md8\퐢%dş.TU11)@ H_l[Vy 0Sn0|[iy_9%$Jk{Ӆ>\!=$FS'|v.р. X/fDߗgf€|UCǿ].V5Y37}hgq?\\_J,M,&M(.a)aSU_EhzY?!̯/2+RIHY. Ddyt |î}{r|^IL z#Th!Dz9cL1o4!IdT*u\FEƴN/"b,i8i-pX+R R*jeI;2 1om'XńK2!XUYIYe_78&+ VIDĒVZœ*u7\'n/ur>%V: iqaŻMV <~C $9q(cwzxN#ů~|N4[z?_~~jTM!7lowld;7+~3MeH)ˣoB{pưW\8(i+n߈$h{^o ;tɢFU'%Wh0"ddHxHGG&[Feё5pr;0O”Fw.sw]֝zF_eFrs>Z" ) +NtdoR*guSF_ @,?jQ9Jų1:"G`1@Nx6Yx2:[7?E|]qWQk%3&WMs̝IّĈ "PMZaiOފg"^;iˌ*3ش+a#lH&i\ݎm\ZdT3kf aGxfR{\+$gBr8y~dfLlzhdF!{xLqfyX<5VxA.aAʝjMDWk5.q#&Uv^m4p-Bsnq,͒aEk۫McjEJx9 Iʌ"o3 Bd32j21WlT֑;5SjJ CQ 8Gݣ4!ָ΋@yymEg2eruT[%xKƨݫ.{Q'׃A޾Eʍ5(N[)Y\%Ni8W7`>ɴ ' (6$G]o͕-cF p7Ot?ݦdǂG+< 0K4b洫Y/F>NxW1Z~I';G*|cɯf2~9dvhvvу3Dgz<$XP}79&[[\3QWVALˊ0poG7q14yLB_;`y(ư3&P_pX6]n<%\,||-R 9 k]@Pm} qVh gV B ~kEk,Ћ/ ?7_Қi/9'[!y9OT+}(PoG~|_?A l~,Y/n.k$4wW9٘ɣa2:-U}]e/?˓,2~Dk}6|LJo %OA9ojY!GMח@>3dgxz%Sx>&5zGǚg i:?sr=Okw5} HR QrMLZ}GgWUr`fzi,GЩߺ7^qWz5뼯Ee^}V096&NxEu-ޅ+\q!LQNbR2jrU7aY1GvbY b9]ͪtC$-^α<зyxC(~'@OF^˘R-Ke:]]rQT¥I>ry' ELՎNnX;9vAtv;^&TiBև|"IjVg(4n-Ͽ=œ[ V' =2}n{d $3 F r,rK O|zS ݩ!n DŽϯffɃS?8X -gz0 ?6G-RJxc3aph,ZF TI;P,Y3h__t.P؟^]K.S$/ zn@?n˾ 53_*o fXk@:~@\1kw,) ICɒ>Kʔ 'g^.h]fѺ̢uYza2+b7 DWAzcRqg"p׉::ޒ*Z^);hICmU,@j#WW5eY. ΀PZHbPu\UHʍ];k"1\wWN2(CVnY-mȴp ݟ$}}gqlTcl[LziK*1+n t$eQI2Q9xRd VsF;#$>X|\ޝHm7bC ňzXFHh`I$qiOX 04#F'=J;F)FfyL$kWE9м $YRJ@񹝽HNjs&$w4 (-ۼpOW{r]*/ՅvU džAi9(0L//N;FoR}{^-WEgoL2_n3w;4|~K; \0+EB}_}>i~_ՄEf >[onH$1NXx2֋LXk!A)e$^Eہ38 `A__t_vCAp#bOowg mPR_@s؏?FlPZI.M=Ztwפ [zy+xL uZ&ZXߵ+%ҺѦv.)n$RY˯iTX˯ EL-S[څ#jN9hcFi/Z\D)#HWZzڵ"ir{ QƆ\юma}b"8 ")Q M[dL  N]e$Fν)aO- LsSJFa?ʚe*R3d^φ\$59Ͻ4k-_3XPK/&qMIHsIo NlT0 {V*xweSDqogdhz*w=Icj.YCS9NrMbq9{ꜽҼ%}MagMxv2gO1irHZUryHz|k,T9->4}:Iz0y/Bn$)rhϢeQ+4g=px`HLq9> sjW'Q<4Rp{H}J V%|XWEi@ ǓL:kXUn+H;ayP?m{&K6{nVvÜ` $ZwQ.DI+-7PiRrٺwuql%YrgZUzpɺ%),+!q m%{zv$ h$ ӍFti4=.1JIUr|TNH#%X|7r1ӥ\"Z}[g8ŽӼ' t.A˹pAJ2"rNbOw]h%,E9:{-_9ƠA32(#1Lެ@R* o#.{F6oiY ;ml;6?姒Åϳ7YlPy5y1Ps DηbzΜG9'eAlC'W@»S>r[)b*PIQy2=l9@m،H{M;3!Yvl! :^[N]+B3HOC%[PsʹT)?U=HE% 3B@?R*˭4] EΦNJ;PB2z~msz@$6hi١TUڙjAI`0A ƙDZ+S1|BNpWi9!Ujv}8ݮ蒉ЩWoi)uqwH]/@ trޥ̽=F|>/s of<WWQ[G'7W'\r/ٚ_?_ rUESPMl/*h-fAwcЖM H{*Ȇm|Y58_d8&޿)*AQ=O,pu{~Q0RkG~W/X^ici*W+f0{/0`܋;T!zUEK"ڽ4JeP7ʪAYGzQL7!6S/SKt`j҈:I8|g]\ڸWqU|'{ %=ɤ^#C=NV{< Wٛ[F1.oU~Y"KG'A! mqfQwY<^>P PB=rgūϴ_v*3He20y!VXjrQ'Z1u6umF|2:`y~ `$  r.* i ?Uԧ݁51]|]b Pk TP8׌@xTyH$OhJ QPK ^POYuZ%b v{߬{[}CRO3fW>E5q|g»Ӆ\k2fUw[VG>r릏s[VbrFZuں6:㩮~ڢDV rcn6*TB.1Vv(hFxV lԸe@p10ji!3S|U6Dg~wJ~Jdj3%?٨56ɂnbO֠DJ% Ljo*Id1T@eep M,r zs֡vEh=(WRK' ^V'> Nz `1 @FC7ZZJ o8GC:b3m:Ղ6&<ۇ|&ZdSVޭ]j& QxMhHao) yKI}~2ƜXLc jiK^5㼻^Ӝ<'|5>:Fw>"A}([UJ;V%|b JCkJtCk{Mءejbvjdݴ3{ <9)O/Z֨joj2 .K7 ^}9]Ļ:zz;1SĽ5sDV:֦|ϼa!YIk-[:y,7"Rb-fw+ tJƻ/?:Z%ޭ0>ӻa!߸v)Fh H[2IZC3E[?uWTNɺb&tv$&(bZS{boBKĔ(8 s\h掇j|! ƸUPT-Ysf^y->^x2Yjh(_O'4xizPD5 3V`PUrԳ c~+C\''j HJ<{j6@^tѴ0:>G 9wz+t}v8Ni`;s>!D*#B8&J!@P{`S;Ombt?W۴(Y^lk(x]Y|wDB;?=P,m(>5Kl>352gJ !`~ QiHT^ L E t`JߗC_AI HHԖk= $fQfs^v2 ܲPNG"qA=9l~]x |]v`K)= epCQǸ^y"ԛ!il[4KO6N[J Q@x1}Ȇx]B'm/H"XD `"N޳9 qO/gD}k5GJ{t`{3۶f JZd2сfo=!a zynbLcFXo},vSy(r$YbkI#ri_nN'i!(d}풛gFo>n䱛9C7/;hfe "MiJi=a^az(Q9hø2@fhUiJ#9n֝鯏:n#-h>`bzirs22znVu-x?kIwp[Ϗj\/O,}F !L G?9B~uX9U^tjUs{zLS|^!JTĠx۳OJT@"^-NqI9*OA@%XdPZX >((Զѣn1mQy3OPKx% NMc*g|mQ]E IBAyQFGÄɀ~ԚXUGŒP /Ŋy T!(s[m3XSsR9i( ;f+c2)tpU PeeBŜּBhkF)`Rx&(ZF# xcP/( D^7/T1jH0hK,X ϹZ!dDwR(j}ea O錿~p}Zo/1Ky 2:$o Ez=N<ڍBJC(1/ޠ{M FT kg%Ӧugsۺ?4B -ʈCJ{UF\zS$ޒJ1B7t 2O*$ (IL.ur]fF@֐EI L6eoOR2,I.VU+JimҲuwn*089(EH1p6JI*Of>*' p *h<#!RQS1N74cR NVݯ&tkEq9 *n,cΦ9W^AQ,1-5SiSAki!TfڔǾ3moDl3:ըb˻!V| ޭ)FﶿH1锛Ѽ[z>,723aq@ɔO=t,59fjm82/Z652nǔ[Pڻ2vLro WC\7}Ϸ's+e8O#E.LvF#VK=©BNHSRsG{RrTs,ѿ(}q|QP9DhY8ozytƉ{ -8)̔ypopښ:U6,wPMb6Ӯ͙'|2 |s Z$xH [!F} t0Jԗn'+pBSiQ-& r*ą*zew %Jmg ʏ)or }/R֭4=HA:MiAژKe; VIR)4'myABSR@RfBo&^dr qVVqJs1dXY9Bmpre$E')u1:ZìHV2+y%eVQX-B*/ uAItQ(7xԠ=9\2@ Vڜ*iK$3v4hQ&\ԌkC5Wd+Y+j5w(Sb+7M%1(#4ڹ #&fV'1M|6[tymi){wy8hL1xQZ~$dXHQ{Hm\RrwwOS-O%xVw~x}sX\E|m,.o}\,._Pr̨>&Z \ا#R?o0H ,t^3O9w|Df=[V!xu2ف{f[nc]aE`<b<?9w ꇡr sdDXtZϢ%ƒ 0$qmqhfj*Q#YEAY{gW/#Y@=_ ΀.=1=h%nsN5ú"1AW;:s3 T%%68eX*DT[гyKRVf[o3-J$gXhzM!T J@bZe#!(H<.#P%%vrv X"o;9( NeYEc{lw)JݒZ6[nIcحfbU:j;8#KMR>IE!hR*)U0RK0MsB x[](vmA2[x5ˬQˆ]hKv7?4T\/rg%[v~z׵Q,ݽΚVi2fn(S&e@27%]7KvȬYZfV+]J ƛ ɼ)irrxv-XV;/؅?_s;B3<0^SlLF)iArfTf6@tGmrp<WS僣@52>gR2 @KyhQ%>_вG2Z JjCfOKddVεj y&ZaSZ)6FQ ǵ5xkz ֭Z>ˆU;P{<p#ZCx kpI@(6rUڰ -8RHwaCP>uHpu,X F=F~qnT f4jQ9 ˱Ԓͬz朢Gh,ik VnP%Ғ|Z"t?G@ThE(_<U^>_68.ű:.I ۵@ߘZ8ET*9WHI6 cPSoUh5*ND)qqSFߪ\wp;Pd{!a(+ۊ[^{Z}xjm Ɏ?g;sm;n,yN o8G4ƨFGgkm4 v*2w Ĉ p/W2 ="K4?O3;$ u@] \?l9,zEPƈFdQ 9) ,0 ;D>8 C wi >C&@g*n7j  DŽ0pLjWh7e{ZITOaVd:鍵&q& w C 8g7X y)f%j)RjB13VAtjZ,5=]HH{Fm)=!b_'X8[ +.5Ώ*@TS!PUO >[[#Ml-Tߠ41I%t91=y&{.4g\ f2,㙘rm'6'/Ldi#3︟ABjՏ%yi sxnJwO{5hl[ G{j==Y#!k!=ʐ>e-dHlPY@yb_#tI.i+P031μ"o"-0YOnbeHZK- 1DZjz%w/\PbBBCK9A^oBūpdfO9ƥ 6'Ҧ) eziOQrj%rp,Ph\JۭPwZsZ{*5-X ^AIm}JhVWKcCa)Mtwo/ HfSDٖ1Sܣu ~[3&̈x2'ciٔDwbKR(*jp8F&V]v\:";dxM $UaFRsxEw/Ai/O-rYrߏB FP>4 9G';GiL*9QKVKH+ ^Ԋ%}-|=_wPn R E6x6ijC|g`/?Vذy%=h V3TS7 4϶pF#"{RF7%Bo2t\ b@An/*F0ڏ:}=/ly'SL[3ѓĽKÍe)L'1f֊Er0O[&Xk˱dΪBK+2[(ye#Ցb`k"mXc&뻇UΗVG+Pc c u ve*cN Q'LߙVBhheLWe]t+% >j$ MxA~<KZƮW"n?VZPz%6aR32I4.8)|b+ CDZ%c%3lb%=1ANNu~WvoomzHC[8~E!sg%ѪBMqtgGH NnLSֵy1SKj|V)y^)2;}k60';ş}5 ZGyx#y &Pz1ʱ.w[._X][]]eh.鲗,i9ت+&{FG[4 ̱}4bӑo *&^\=|Fn>9am}WN?7DxFf>]Q.;|;Ӌ6h9m2' t5Ϩâ{ZC=ZӒfl&/P NQLX܌]fD (8O> opuJ+VAnOi*-NC7zxzȡTLf>v 8U`MDgE͌mjB#:|\zv{vsw,³w[եC>SDqa9@X┋9r'VBn/쯩KVH ^^ȩtp鮞?S]=&W#ӿx|iϗ%ϫ9b9 mIYa@%[5ѓ}}^B'L:aЖcF:C=*ȳaҍ V֐es/)K6D G#o>i OΚ ဥaLj#b?R yuA~^ya)gT]mo#7+ yNp. AIhv<"-YnI2[nF{j*d0Dv_ѭ>RP SN4}Tǁ֫)*xPnb6^[=e@ZZT($\KJ;?  4(֪`Ā#D&VL"Z^shV/`q5JVTm<8sW*C \s DhL! M-p#%d(4e-{ݕ[XWWg#zz!Q`ze } .0CIP$-cbaG^*Ǯ%q , nwvA *=ݷEFE6:NQp[p:{(;ɠZьl(Y'XgGa-ASz2J3ٻilw@nL24kmY8NյwCX]YrC&AD< -1q{#b>1+ktTw8t8ǶǕȡ/68ڠڮE.H r6!vUEUUeĨ%#Ey!>߽pDpx|uX̶\"eROiqyrqqG? ×Y|!uqspgOTw[@ӍTqA0 D>xcսjէ1|<>pAO5/N4=nt^ h`HG3+0>0Gіm|J؟bYd)a#`]ppHֶj7NzNy4ҵfW>'Y||βez@e ٗY q{*l8+ΜrgtL CY94?ζpEʞv"eOzjuwSӐM:؍?K׍Sd} VSm~ztmc1Q՘qZ J?3SزjKIUKsc=Jyۑalj @N~H?D(PvDr8}PL`h:rO~\}nOf?)O&)[$/uVXWBIԻ P ͩT˭ + JJHr7[~:~<BP+229kXcAߏpNU'Uߙ]2 >ƽnBqqञC~ԆmބrHO8PH>ӎP3#rbB FżQ@F_%ra!oD3l(vSѻbb:mtnGQrF½R:-,+va\VBXMUz`P\BAС,7KCJ]!JXUdaE 'dފB"(3NRN',(2k&ʀ`p.V1ꢖ3@e-Q R溸 2];S0z\*Ƙ2ۊ: /Rڎ1x f[Ke 5 E^2)JZ(s" ^J?ŖgNDEHY?Dp1*|jEU\ џԳvOe 9^jÚ>i%&"3J 0T~fB!R޾0Xxل_}$崴-JÔsӾ|:y2Uʾ)jHݺ%A]ׂ 1PԽxH^th'&SK{30# j9?~Nh'u@iz܀ w f_>QS ǷW3ƈj3AYYad(]Qp`q-|e4i4uNx*UD7D9JBlP(GQLIGI B}Hѳ^dS#jRDtGB^8pJ.p[8e/;%>޻J^[+]nn)_ ͤ|$kckZzȍ2@`hx+ӛ^ u'kG98wZ*$G#,@rKtHzAb{ -S$M -Ni S|3]3]mw/MEI[@Ĵv(l-ܮϲܮ˧Gb}qkTH*Up/d41ʬ]W;O֟QlQ6:Rsv0àȠ[^ylG] '>e=`PS~ۉ̛}kbO1OJяā;UIR^9^][h>G1M"&G !#&*=JŃ;`dIUVn+VO1ݨak Mtk*IK(Z CZPjbH͍P8,bVD::uڞC3Mp9wi!Fm!\ؠ\oB`ƒTv Pk}m?-XJX4>K Mž?j)^2IEBᨧfK*ТB J#EEP肂ΒʬK2QM%+=`8Ef-V)K^hQtVTp1 6ʳïXr1}6%V񜭒ZJ9*e`aսrz[<~aNNT>dz׹_i{ٞ~A\|H x  l!M[r*Qo97Tޥ#~(-%8_Zu3m|uQʌmh|Շqͱ)Na \Һ3EYΔ-4vrhuC`}p"Kjaڪm?:֖i݇ʴ=pqdfn8>(c% |i={0:궼]R)glmy8P˾ uvU7CAYQd*UWגI`y/G91s0Ącbu)$K }QHq ;<)!y,n' F5?__x* zzQS(&8r/ QbO};f$$h!jɧVO$Ô?'/*ݲ](\.$/4QZ`@9SBS \#H R lX ԯ_]0 >PRc4 $f6h;? $-eZ,TD9ifZO|?}:z :X9}$i3Mfmm={;6.rw|AIm) `S.G~ˑ_rWU5"g&c`r_ >6&1J8#(+ι$>_+X~e˯W\^.U^\s4/e,0e<^ Hh@"*^J쐹vq:mrdQhS~ ZփP 0܉l 580!|)pA5u,_CK86_`sh= eVZʓk` -@n-!L -N텤<Ϋ g#v]jB\B58{OcH /! mi 7~Rgp!E6(#RxӌG̼@njnZHm & ƕŎDL1'AqAp(ϘAkʼn&j%RHV`$ђ$ -I*.jȼv oXcr%yXNI敪P1ނY|wKszȀvʴ-&:؄ʀ:Ɩ.3#)VN]j0BKQg-)ZY9h{3ǐmy{62 BDmvSMA,Ǩ Ab2A^ئ8nJySHQ+m$GEN/F6eX4aэ}Adea]ۖf1}-H塴ǀrTf`0R2fo<`:qo*C">.9D ]J}2{BFT^k n|ۯӁ$iGg#e.VgA#Ώ_t+4娘F뤚r z ˬʊV2Db;1V|P֟䗋"t}uE W%#Z+W)m4"*{:}ϟqY~EZ. OߵA1Cnj} \s&wA/ö38mCQh㐲Zj=yÿA#zCzƉiYx7c̐cZ wiHT#ɮ- u'hd|r 5"],ZG<;y)y5~2U>^rqG%>a42ŇhB1qhת9fg|C,p0$e](,r0a-%}[뽢 yk{) ߄u;UȩI纸dv>O#ֳk19CjN93 Lh=[;{@2Ҵؼ}8'lq] N>?Z&Yzac, ^_] p8B3/QPĊ(J1*PpL^,b&89XAv!o{p*'|( Vj[q MT1XMB} $8iV0:&Ք()Brf@V9T{0<0K|,X92$Ԛֵ2Tک-Mp@jԮdK gScE5,}Pc엁(0ipŝ;p5MU`<MII_WFŊ;d>]_hYAEiE9*G4ZW XPQi%TZ;EdN)* g~xGB m'zh k4pM; $,ܪ]Zo@B ˎ|3.cRk()y7RZv$-R,L[E&3F+މk#mH"g ̆9!1Y0.1Ti^kEEɓT)FM#)r>Y{_V|;/M@ 1"29m|H3r懵,F\m/׶YZdtRQ!aݸX{ ŭB\]?!.,L9O oP݇I i 狫iabY@K@ Rh&aF`O@d[F*SjM4FEAF\ڜ7ݎq8D+ `/ M+c:YF)W4 >a.t?<&d|~mRf j}TX~xE %>i¯_—u᲏k59ryz}z,jn}_jen l7\!}tT+bil8~?]Sxq-$u/ n >_GS+rjDt P1pp(X47ٴǥՎFslwpij AŒy%@L4_8P=P4{{&||wN]Zs5y2-T>dl KӜcwB(ތ,5r:Ba, Ba2ҠE0}<6GCmtzЯ3z_ϬT-(\ HXgKSsJ%{]!+ `A%p{!{42{4!6p~<ƀ"8ao*B5n-FP6du iT;=G|j,4sP3Ƨ!mTHoJDiϾ,_1bUÜ2scնodj=VWi]Y.9b@єp 5*/gU&g n M^P0j% L8]!gt,GdtґQJ1CU*-%)ñnq!:HwtՂk"4~pf624G%h.ΐs!]sQM^g,mhWL$kyQzEG3= F֪yZ[-q{E큼b9HwbL2(ޖj:kRIG:**b"n@N֚g6t wLinwhŃc4o +^*}?@ڕ&)E>>6ᶃ X9;G“*6 ɴ+6p zLڽkdڞn Pkcb}a#wξ f3rSR]gl"WL:(Xq^1 ++: F5 W|)?m跪ZD$)I$2·v~$+O$ kϜ4("eD%_T(STPƝD2:O0sRDCSm}^{jqUJJuP4=ZLAyRd58 A˨  D\t {1Hş}H է,oYd)VVM+VhM./(.(ԬR BD g$-[H)BD RFC(/,S]X862|HwA %aƞfP4uʌ3"\?뺠L%W6*YԨD݁2:7) P@Lt*^O)#1>e^݋C)sTdiǓ7&&|׼ピ6'gx-1dSD{9qj=S`bYk\6T8/Y 8Q3!t 'ndߜ :N˹dj lr'kf~(|K5R!Y?9Cc@Gs7@ Gp^0fl9WD`qa@5}NYE78~iu9c?Wp0=q-=I]y d=F~ ohRs ҸKʔR⭅VLw$AvLEu$scaJTF16HQiJT r3l 0 3gQe:[[AdLPU*,44rV H(U^F;-J 5UR]:%!!%NI :%vJZbM ;M%BI SҊK\-u3f9ᆵU `4K0k0}T jfo's0q( ~lmqoڀ\O:̖W"~Cex (rjI"j]hEkHes֜ A cmUg9SNU/<zIh :f5I֘[d-.eW=n{iidizZoW]B+*䛡UqڿsLkGS0: %x2ρ)"Ӏx% t@ >2Y.!N[B g2So2}9ULuW:7m8 )`̥AߘABy*%nί+kY6emYcN>?[8ܧ"24r^M+f<~sڒĘ9տ>oMƓ޸Јu,V9 -wj :C)I!6 ًl~\L$/\2%ҙmC^Byؐ^{z6B*Z '$3 ̭B(l|ʵ QiV(*[*,jध%W!2K| N :0R4:d'' b)UR $,Nd-.d98$vōWB'3Ŕ alNS+;E [D( WU>.#kA  -6?/4I$JjUU2X$ abRLq+p#ND+Qg'"e֡Mjj' ϸ%ަ `+ma^3pld^ZDeiwߓ65^A*?3!džn?5%!õd[KA>I6e0?fű&-~`!o@^*/!G`Ido[{AY.2sŲP1_omFEݞl!Hn&_.fgh,^&ߏl6lidJ,>EHV=+e=:5蜵l4mk?,Kq>|x,U?ГܟGo5>3}J31=.>ezq"Ò *E[q鴷D|szYÖvG}UꛆFf5Y-I&m;%7.%іQw1JnHQ*Rɚ*U& ueͪ +o%^͏ W8Z$:m&Wժ􍎧@> Y+"{vnkuUt嚊Ę@ۊ<<޽>7~|y70q i궷\8-\r-ifB{=^Gݧp*#J,fv/lӹλ8VTӡwvQ. d p*$4}ntxK*&CqVqᬹUsV<ߑ Ƥ6Q."wr~/G+YgQڛ8@u;?Lg =߭tn:_tM(G,H/`BA^i>.g$g?fo (EsEi軇h2r%EvhM|B" ʿ"bIJTιVX[U"̰"!Ը/XRHIZȓ{9I1g5s:J ш9Y'Ƭ,($pzzJڸESBxdd.IDHf(gn 3siR͔`xk) ^/,̋fǻb^ňqM^Rk!IQ!x-4(WnCj/KGM]M 2uQsLn&OSv߽>5w3re5yA:?-:i[Aj /vZs޸łP< /Z쫥:1Z\0o1{9hg* ͐3Is*(ZWrúQ*h[*1S':1 ̺%4ֺ51C|&).nn<*ܲL)AUZeB%XbbXKX:k)A1N: wȆC~/ BUپ{U 5Oe2Mk3 b2H> _p9AogR>烷?E=>||lgG¦Jppr,TKEtFd #I=| në`AluUhYoAI4\}#?Ͽ>8vWnv7H !'`=E~1IBo 8s[h6O̶ 4ߖGs?Y*s5|'$ i>t8dǏJ16+y bds M:-r>Z};xˆYs{ҳOW7_f,Í:>Y}bwKJ.*Bǭo&⫏'wŽGe ma` P<\ % r@mCLe`ѝRz*F'=c,Z-]';t}2`+>z 9?\YvGr+?_|;kr5-a}كzed/4+s;-hC[h_.`mmK ҶM^ Hmˣhj{_^x %9CVVLeH.PW;p◧@eR+O_]+˽2w=T;is' nƔUrt1I c| 7PqkqcbC℧aR5S(mk9]N|CR~&Q. ]t ω/j5MB$݇:-[wb4UһF)mվ+@M|( 9FN)kY5hB8ty}w*l"9 n2gΆg 2gP!ɶ˥ep]!Ցot(~ ?-[U)21 VHe EDuGIk:H{(g]qϮBԻMzywJ;ĈI^cTIh:.9뷡#̛flilʹa>9$$g?]FX敔sɕ[`$˱rQk\K1 H +לd6cW4<>Ia:VH!! 78w#ԐCn ĭMުܵQrJU0ׂ@Pȼ[~uDUf_\$j{7!8XʩZfLHⰓ1nLRE)( ZiZbKVv T9 !lyv} ӍD`@@ !5ָ3(ghu` RpRF AfknCjIXwS1!7Mש7 mbMS &oX7HbELGv}CPm6ۉ)>?N$\ۘNfu鿘Q̊jpZpsVK2߸B2\ !͐OChf$ ]-ߚ tb6yW@fvqNi (W1w˯tLN4<̎/ؙU89g}c> !'8Fw9r8@*n\!nUu@JH6QS6 Bjl.@\M oWނݙL@+u'ȂӾ0OQ%Q9Jizr[O!> X&Ԓ*b$((\U;cG(+,cX#*9Tt cQN5n%0C3 S|p'DB0uUCB/V̷ܴ^oGVHxeÁOnn!Yv}UfLGhcP̼6¸X06SBrd39EB"ܙdٜ 75n3sK aUdPX if) + QZo@"!cׂ)NuQI &,ɩSH0k );Eau` .5*8:855V %ϘB@)r[Zs8Hv9Vɔdǹy3L Ɣ`VQ#,r?r)|j+d\Tt uhӫeQ۫ 'iDUslw:`t2n˪Qh|Be5D**C~2J0 /b3Z?v b{*l1-[eޙt{ECa!|{fd:[ZG#Boόt*x OG !FUJ~zoTJv?|z@nR֍33"8U nsN1-,P<ظPFBYe[Yn`7V8`9ڨ'e'a],fOonn0H'GOJzGy(cGkf/?mW3O}SnMq1e`^]3A?z"(j[y)Ov:fPM (7j]|yKûU3CaW3el*bC3;Ok#P* tm,udS$U++#浚.,i8ӦK$0ŚML;ުjEg4_ԿU 9o L!-3(pNVB w|_fp̄Ӥ`ݗZ4JkkN.p-Fjn 5Jmk_bh7PZK,t$dGs..GsjsA1u| Gqg],^ViE5.)CֱopL!*c6< s^ |~0^dr{zU.ӑ[/ο%ͽ_6~es.?Y㖜wxn~.X;Li~@'\ =CV( 82GZ;WP zst :F7qU}Ƌ觹}_X Wϓ"Zv9,?)]Ɵ֡{wslZ}: g'÷Q"|&󩯮.~!'~Jcv|tJ˽ZV>Haȉr@R6\N{yݓ@YP2Z~qlh+@dj^0"κߗ쁮KE!#HIY-=x-N]M.jyKPmXo:Hԁa0Vwط }mՍ쐾 SRG94]ZpB7eh7؛tLJ(|i={L)U#P:Fב& (C5="Rhll˙70 kfXl/.mumݢAWI$ FhR` >K0X\wرjAr=kBݜӗf+7 R< G`VLGͰ|&x3ר_,//>.vM82"+VV:Dm 3§<우kO{@b\S*;[șeX~KF`{t;K 7 n3DI@j/lK*HU/%-kno r5AjzVе%gГ6hبV=~ٌJ2;-y c~ )B󝊸t]ø\(=;w>/?RaIό>j;4~ȏJXa}wO& l;l>oɕY [[%h8tPTcX{lqr/\xMu]kOysQ(\p*D+}l]\t&mq;-/s ,6K efkw@¸?җ=.?{=#؟@[y_dpvbLRn w/@ lp{؛'ʧ.%D߹=:]Y#t&A>Li56ۡ5?apK3 [?{݈"p@1: J&1|2F#\8EȻj3F L8b.IC_&"h'_v1mF5CTpbK+짃=7Lh!Փ!.+gOm:q*|߹/8/ 5tr#C`/aob;UYaŪI;gwndfaZ t*P|%<_N@^Z!偕e1K͹TgԊ~Я%*_3_fY<ۦpwErc>NW MYG:o޹$:.>;nŴIfiT}v|7xiW}NUp?X?QPWi}%QV %֒_wj 7(J{L>m}c6Yi~aJ:9c_Y}y2,t-r< NIw99xRh{wů/Bؓa-r]ZՕ4ૺmU(M*D8'IG!I}8?g-S/BYR=SIj[|̜NgH؉9dhD c:V!O$QPr.J͹SbSXk4Ǡ=5 uLt_ 6HhשA`H5{509 B%8{nH ܙ vl]e<^yp:hdVYZ]Fmާ,5.;fg߇vxk6O'E}~Mho/#k7+6q@]_7..+\c>vX\]ݬǰ2t^Elj>=wWbVfIwaTFr,Y4l^fFHx@ТnvR5AA?Խ8| g VVPy B][#!e ImwPf'NRήN[-2駏d )*.-9#5~aqhs6wka|Սbҳ7J_Ҽe?rrIڷu"K'3Ξ&ݢڞpמZp٩thri4!8;6[wa&4ϢiEMXRT˚+;:N`dGUdh+%AL֢jXvOb"մh95zXrSDOE`D\6ɩS=} j{>'n1SQ8_a8J㍻dy %Cah:؈Zա:ȓ"p5{&r-)ɯ[͊EczB LYhW)Nwe0ջ `!Dlʲ}[))Sv]Oig1w+g=NnCX+7ўMDRsnC=?V =2Iœ)ۭ *9_~U*IaCNOrD$ '‡ ukc' oE5~-pԼ@B6_Ay/@F(2\/?%Bs=1nBite+Yp6 WMZpjkNWK1c/RaVv0`Z@,R iW9ݺBsp" L>xrD82΄ C099hj u+kqa 6d^5,5{ 0AȡOѿF1 WE)Ǭ> (!͎x|2b5 -,m^&T(P)<7T$|ў= Bwmq|J{TWu Q!X$@Ĉ7#)6!)g8}3ִ [f_].}9 \(3BUD L6|j8+P6IBE=VؤRxNfQ8BRHxEH dNPB[Z1ewGP" 6[4M~O?e>q#b-ծ`6|y\~ /B؉(.RoDK+(eu]iee}j 'kg/e{wp6o]P$ٟv.(oƞ*ߦpP8 ۄ#Zlm$r4pL9:vFmspۓS\ZFO`Z/VhGipJhS#T;aS=`5Y1Et] !KVz "o|h-1iY@P^fC@*J:i\2Mc7|1>@2L@t2u'tE JBIfNхr!ҦNTCY9CYCHp\'g)_?ꁡtv`JQs t6ҩLd ێ-yu +lʫk'^;0م.xup$cyPH-PgEʐDjR5Y_cvO:U'~0gqU`70O󳎏K2g8;Sl5q0hC$2f)HE|XىYcQl)FkYa<&GY#DX FKNPYTdV'm1fBs8\7N3yw'|(ȏ2[ dk AtW+KoyKsWr$woԛn+K|_SZ-E=vBv? F?}ZN[z?,>}pO՛?Bk2_-ݼ-o-E׫~Uo8OggV%lzdHk#NÃ/~?ؼ-{-"JAPC 8h$1^ y㫑#6Ob._".ڰׇ?2~=1aKPWNm>]f/1ՎWw-Y}"_Aفcu峍O3XZӹ튿ə=r]՗74kD.9EMOyB $ mk8J9z~-evU6C9eN8ӫ#J0s6AQg⢜aWLh.^$9u7.Pٓ."&tخPD}[gӆ6!k#!c:(m3suHBAja9l1#‚gS[nQ=a>iWzhMe_Ϛ?XAнeCW}72ʟDV+,o5{T}0_15H=׷e O.Zdp?Gs+[xlv,3$a&a OLn;nW7}x=aҪzh潘!㘇1)Uv Q}z謁cj+6Zl0|^> ,QW?w_ (3faMX+Sei:7eido=Ksda 8aԕKvGvKoj~٣SwG͎Dﳻ^ Ri10Ga=Vqٛ75pX 3`}cej ~O )4tʈgF?nwF?u˃3XL{s[wm ( zgкACSF im|F6UNh)GJ&\Zww#5"_%)sߝ }<ޟqai7{+xuϹ&]gfԅܖ+E͓Rt$o]cuQO:'xqpn BQ^Z҅LRPsˀ57nkYhJa5SuVYy)锵AڕT)Y4NzYkW-oXamr-M-/玄Ež,}=ee #`_j?Pt~ D(,snXh՗14c,cxdzB93KӡG-[cH=uAh eہ;L6Rliװ&36xS6ٖRʎj#Q!D5F S8bY'Ҳ%); K߲h#^.ċd2]]ƹsG$ }{VE]4V8ǖ )I|a3I簎蝉"gs;/>NX཯4bqT9+ &' 4ǢTk `ܮbA)O"|Ց"%dv{˔9FA@㎯/Q 2leQbb} ,B&I<(%!W,Q+k%C%XCb0Q'fXsȋc1> H:M v#Pڎm)׊R.$ "xJ""k, lrꏑ5'G5 IDq)0뢑orζP)I vxRl:Ĭ]I e4!{Ԡؓ 5bvkBQ*G(I#f*X9sU]+uD.#PTfG{ +6A$U$ƬgQ Hj)T,!rcVA^ uHJl6:gbcH]|Ж& +lRHދ4y 㳫 -/@7t9MC {;iAfFߵ$:8UJA3yS`l5K*Fw].SmEܰ¦N`uE\,6໠?ΐ.jK/6gu~Ar:uո\N}:LzC5̖N'_vFJS>I̿5o-Ku-` f>X!!*ڇI2G'K. %kg4&ڐfko w)MJ=/''~ZccdVz[=+h hśP{y_/?_ꬽɇ5p9|pp5uZpޮ_{{?R+ϟr`NխU7?+g_WVe=緇O|+n \Ai{Su?p,2ZWd,Aț8ug GQx*Ѣ6gB_(d RFA1Q5aУ 9zHBL~N8m6<13vVґG)PV%`Tmxt0 sHE6.niXaP-ԯ3 Yj =9 NZƎEҶי6WGM:lkZY8ڠ/gW-,g|g0]|1ɘR;D_irVA~P׫1)GkCc〞t+kq<> %f9n3~ꭰdQjv-uN܉gY۸9UPx?^GrWO02cdfHۊ7~JQC'bC n *+nE f3I[)ٳDe/s'KF8ޤ$IyI.=lcm6*[-$5ύ!g1"Kۘ [Ԣ42xHPtTH미KM^u+fډÒ)MaıiBRzpxGX㏘}@mRkƫ]p05l(Md-<(C,r d%`-k}w|QRHP\Xo_/rHX`=.=\ʎ*Z&ݾ&쨢Ԓ6_~j@:a;RvőmrP.C-O \j:r]a-jpjpph}5!Z-ZBѯ+Zն,8}G΍EkT,^n (䪈}5yV[߫™(r#7 y1 |]z`QUGMn&&MP.jR;r3ntn.2ZXPq lYXiA !.c5ҨUDHԆ fzRZyMxtCe܁ Z2pڬ/$sPs!XV#H4%8Ĕz@7P"0!KPP3 BooR]2L{n1Qۊ$KEGmo\*;r4[vq ?Xȏ^*pjC%#)lJ2Z`ĀmQ;+=\`Hq EVˠ]Ziz/i^PS%O="y=a*W,촟|*fA]RvyxYǍYSNVg$IuгWQhZMicF ƸD2 5{عoY=4p*6 WVVӭ,hоdZrq%e;Seer3ŲOX}Q.%4$K "8~=7k^zK_r~Ė{Ȗ{Ȗ{Ȗ{m\$ hJ8xXEK!g\+_XN7.h9Q'Dc5r8 $o: k[g0лy1lp̀Uafah40a8.{*"znV {+Ѝ_ tU޽\67r]9f_/~Wݔҧ+خoL+ ,4zWZ'RRYWQ`t^VR@\,x FHKp'T$+! R\T ޔ D4aěQZY&7A2 v!NO0:4{,p$0tHQfx: Ɛ[N/:zN1] hb{6bM,.TNh޶BB7O"ق/Ry$Lъ< ~H#d#WϾ]T?' өVL9Dj`dW‰ߟ1x-?$EN{N IFi͓TP@'tr؟\^z> ` q1qp[ŅH Vd (P/۩g++٧j81Ȋjf4vw1@ 6ԝ]nՑv \jZ_[Xs.J_ݥ]aٻyq _{_ {m0+} Ujbre bjtg>1>Ko|:15B];O+c& nhMgяEmvqvM{*xViۉs)S}{N qѪtK PmG6Q,DC"!DnQ)""QB=|?\hishN=N@CӠ.``&>d$°Q0qòbò 덡x{ z Iիh80b`3D%_Ma3i;o>2Q26ÇٻwEoVIFL? w繇1u! 8>t; Ӫܸ^:ֽ-Zvf<,w'~=]'Яճa5cA'7߿xHy=Gq|iZg{zpUmlǕ/χoρ׿Kg?yzi\Uc/ + XҍK?n]H?TO|S@U=l&fs@m8 w괯Pvzsq62x)6~|XFGͷ7:HC3( f'i/'-+ !qp۠,=M]8,~Y ?¢6xpz9JSij>~R\ +PdzCoކwFO#ލNЙ:G5OL|B{Ou LJIr`{ f߼'Xi_9hP[0F/%vkצOhq}薇#s!-|_~wLDs*A&c#5Pcw:!: ڧZ2~jNtKYyM45A[c)/ŋ \yІR]Vp*ny2L+ +% άvxt״$`ˆ3ɟ/ q&f ךLƫ]-xU׌DEӟq^Rk zS+;T Wps+TdTLZ{_1ҩ^n!t9ԻPKX\zvwOC^r㰞Tg^ wnp>b6ysyLGA/z^æR?-_ Wу50c) 5c)Qa*vi2>mq<86T[ "Rgof?alelx}:tğ+$T $TV፤U6 !tk!lRv3*U۩x{jB RiM};>ݒWωF_ʯ=6"} We]li|~8$b z}CD>;}7/L)ͣ 5/cҦp`(fugBBz|0.1VJQCFv.?b6[NH9O8=6=ssb|ũێSOwt)η\B򺬄~-%㜯y'FPSVxgaU~lçgm1₀r֎ljn}S[ 9oJ[¯ZXɩf j+9޹n5 U ز$? .kw?2 Uv;aj~0=qϒeɖep~p撲A5fAyc,JNw½!~± ytk i\f~k_ŹB8NM=L Opf_ۿ yمB^ Aia,D:6jS!WMZWϔ)DC/-<ƕ/^fgzvM? =:-%+qcpJ YՕ9M3p Гd _f7 =s89RBZ b˽F,#0Z|Ԁ(錛Æj~wZ0fHB0pz!_'W C*#UW&gpVyθyeX|rx83b{L2n^q@ e A7ۄA~ܳ Np^pz0%^N6ݘ(ɓ^u]jpШc>޶vgdtǛ0N{ \H_AYbOi%2R+铳ɇ|g6ߺvy+; _x"|Eć }YRV_`$Ngz=H<:{/{o>x7:.QOvOu>$ k- —cz C;fa{QoxL "}^SS_B'G:'we.7}tu\5+)әkC(!W3\-nTǢ1+ϕ| 6$_#; I@0?6[Y=fD2ߗLjt^sЙqgKS`O ҀU@QCr/dɂjƁ'ʧb<5=/~Ѡl;Bzys\//ć_+ldEw!+X&I]}bثN^yUq;.yvoM'Q̏ԧ<4 ^kV۳&:y(͛kTu c_ s*v,!1*C%)T[j1T3.vZ=7xtz-uڞhwILkx;FpaUnKI\Pe;,kʕٶEyOIQ6ޛ}v$}!?j1DSJ_fm.v}72ser-T VV1A|p{ǵx8!EzKqH &V߫S\RdT AaN͟f䍙(闓.OMϜ2jQ:Za )U[tߵcqd c7ԇo6@U,#s+A ҂1F<d<53`pG\?Z*[ > Qb=U%iG>s0V̯ut$Z6SMUIL*S6 ţm'?77_ Frb]OćH*j d#9tSߥѱmPg)hs:dʊ>E}a'p2 1D{BI+v"WA1=vl=i U(Br UJ,[4,.|9W=lcE3xx\:X\DA{l25xu4#Émh۞O =_L%A^Uh/ݟ0Wd2$$gѦ/Fsev"ۣoAD -׳cnC 2'Ԧ1&9YKLÖDb_9,M^/_$rFS PvZrƚk\\*1;ByYmE㢰xigVKd|Om(ifí Zi HL@r +u-qu$3f{{dŌp 12]pYId?`3Q|wǂ!;ڔ;L%'CgrԖy2 ab-m͡T:?ϧl([E pA5+Z⟯#x !'tki5ś.o-mkUT1#S rSdPR,k?{+Q187u<}  lGP *fK֋sAůx׺ƫj0+^) Fp S[{uq8x\礋LXH@8+M`'4RHHW%nǓYo,6^L:[5Q.L}r tl4v49݋O^87I5 thunCؗ`K8ԗ$/E~i^t虿1H$** ~$qqqԯCn),سkܹ`+ J3^?xA򏖾K-uԝ8rϷݿzW$kw -=)l0*T%ɗ~'-n[?>kOΣcoJ.g AC5:c||#g~u?4.s>?N\𮇡خcގ Î5Çpa}hG•Ԃ2YG(x JNH[h.ggЗ9;] !'^/_>-5IT\Aeђ֦ܢ1#[1~qf :OLҒ21BG*7%. u'cw~̄JjťEsZJ(!imDҔlusޕgBWHMFX/JdJ\Iy21SKL[VmÓ,a>Zo6tIlB,6)'U!j%3|Gϩ*wJ!V]fEs/*$J β-> ',/<.Zs>e PV&oC%=ɒVrRۜ/QSz+f/J,L_q|,r4_#:Y޿| jb+)o+mH˝`z,MLbW67se(:%J[kjυJ VEB`4_`t\~_=>ȅM~a}:aO͍(t@G٘ /z"qF0Q =fry8M{A0zwsvs=J@j0`|=`!:&΅4|X?W"-(iLu^`pYq ($R_8`iͭkkz9@Ǿ,NÉݳ7کuDk> H>7)P"= ] Q!ĸ xe1Q ˼5Aߥ˦Xnh4-Z`rSN;Ƚg` q'* m؁&{ֽ}@4 @UH Ũx\#PaG&:*z&aA1{Br4,}`7S8#E"3'eSSWY7[6TWyU'CUJ)]7YjUJ J\; Ӂ w.?Ċqp9$Qg烞(#V^bHuJ"ro4"UW Hn[BMC0z_ R"Ȕ_䘇z T@y1,tW͐ CyW ~qϕçn13͔J 0%`0wJm/'?p؊ݗxՌ*}LF_Dస!j%ʆ@ )!`lk8av Ktg'$JBe:&{3a0Y<ґyD%H 6V u5LJ( q';E,a/vúM&>I/R#X4 z/e^ͤ ;$Z7k'˗By4š]ɀs\- r dBh\Sr7c _&ȠV ){PFD)Bx ra.FLR!ԇ.)radO M*A&~DITP}hjhFvy?Y erKKJD$Ɵ0N,4Hr?j5Վ ߼}.N[v{.2n|8Z|?Q"(]cY^KY~&vڮ1. tZytYܲA⊚:` }U񏿟ppr #뛫~?Y?|ևvjq2AF.t?b^<d&0̛b8pY=E cpn+s;xоShi|\(ëG{U U/mi%&x5ŅM\6?K;XIc;Wz'gF ?B"i949M܄V֖)r#m?7}URH?NʩUv^$ɍqTeMEFgr|l  xyƊƊƊƊ[ ! T+9ˀ 7p?B"s7o"ܹ2we6uؚHMAkA289Og`,B:cgZ>5=-Ua?xm䪳'X`RRIw1o?'g"|;$iA.~r#ڡ$HOm^Ο3*8;6T"LPu_8(Mx2]p 7 7[ebMpndz׌':Bsky<VQ[= ` lx+3*$a܊]5F3, [@q}+[7hXK`$9j7Ż!b\O}687=d Z 52Ր巾>6BwTNr 39_;%S!ŕgqnw(mJlcƗ`:Z%vQ.lm@ߖ8EM V[b)%Ɏl!Ol$8oЬ1 ~4l6#1 9}!JEVFxn-=ѭcp3DgT݊Th$CH-M!isCk %;G;>~>lͦq<Qs\lAٚ.眮C`;@Gk>OWʶ62L[oo1kٵ5aj~t[=urxEfP^|?ƒ={t..O 2Wdi֚r6ZZ9|T=uu8gZ!NA +H|ofɿ~cGU5:)u4d1Iz,bozf6YY\|xVt:#,}C!uMB^'[dIҴ]X8)M + OV{tNtpmV5e7eNV{㛗=l=wmHW~``0HF~5`/;F#L%MKj/J*d{DXG>E ÷|Ǐ M)_r/{w{j\V~k]]-M(c@sL|w- וyO;mG?|x'CԼtf{E%(TdaŢb4m;Q{o~lߚ+2a j66 iۚSaL4襁ЂByoZy(P:htM0v;QX'!N衊8naj4_sŝ t\Qh-n8mDrefk )`P8@dC6_Syh(8؎ n~BgOQ)D-= 9=α>_ %[LD {ҖV͘$|/QW d_rfʹNƳ/8p{]a ,"֎i1߉||hܾ2@Tͺ|jmũ3P@WF;'m=I3>~P"n w }<>,*}<xLB{$}ۗ}V[P exTgzV${vOS?HeKpPa<,lB%||JUjP;)n.%~Q;c!Fґb<ʱ>V1֎-8ݾd~3Qƌ(V== { ۻnAmqZTSlb֫߾_~{jo ߰*6`o\ o<9q<|2,6ˑ5 v<7M\INz,T<1cȀ™8TlHiĕ,䳠:bӰpR^ODBa2"MR MG>g4tX{)QiijkH3@|C@ b`d4IJPqEm k/Z@.V>xr_=.X>[X>]?>--[?wߨ-s:kFժ6rƢ R l'3XV *"PYZ;C{;ynՎ╒^\^?^/Fxi;*A)̕i7CrB?!1nb} k!:ız<OvG?M-CUL_V3)T4Ƙw1K5EH5΍+!U!vZ9&k]W(b!['K0Zkm)# -۟2&EIj{X4P [o3(@4h[jQ5P8%{upZ1(B~+6O">Zout,BYeuۨi'2nuޘVo5P<ҘxzqBS4 DZ!G"%!fQw{\TNGSZM% %Isβ1M$xЦxD HUJؠ/Q5.)7zQ>$@[Eҥ4=ny< vYA>S_HfBhԲnG{awsXRk.0#,'wqWT)-5r5ł~SV+}zJ8.'?̥m%^\&[O~) _O/b3&R|a4s yodX`[$&29*O$Q4YP>4+Z6 K& 2 t\醠T03dOnn _ቕ_ҖC/)oU~Dy\ K1rV~IT~)dr3)^|Ƙd&'Q'\EG ֻԛNcmk[yҵwu7"61Z;JhTZ\Umt t>Cz,v1ůWW栝 11;Y7sl8[\@"P*_&jOˆ`ڨ]oÕkD #-|&M!Tu(i Pr6xR:~*.HNVUm6V^*4j"^kBƅ-\-֍X\o[mJs͹e=SLRl7ƶN67(ZX&ڎ[؄]JX_*#:K@#vh1i9$KLiu1G;$?tz2K0O^$V{:i Ko\7P(_Qa«kyN@ʻ_MEZ^WQ/:MDўozjOID yiS^H<|raЧ7G+$4qIwGb/; Xku'ƠP6[Jּ'!=$l@ߤVۻMII_ /-C^ʢA]`hJYǟ7% F-}hljiT #V:gjQ92:8F.I98 M}Fɞ0z4Ĉ)JB ՍD@Y_Zh ?5.jd)٠(eVVQRn4n"P"}L 9qp֡ZP7*4(Z10:ZJ}=XIWQ#@jjӛ8JK-`&)T{VKmSD MgM&q:kT[DBfsXR ڎ#\) =gƷE{fx1՞d ҖFl>59/gRLܻ({م{/ R2Wh&.v¾Ms;4OC,Q8_.):mfXwbMf9eqEpО6PIBkYƹNfd̮ ?ӵYҖG=fdv6̮2|NfY?8͘D%Q-HͲh~!?mdGɳb=nǒNS;0)eC>li=he?C8Ա/VskYWliT\]殚1kZ}W7b/?ovm~رƯMG]b2YWv=# a-_feD<_k_׌[VvPxeƆ jP{r̮(k/ǘ&",p\4^i3x^YmnӘz;g 4MEs%'8(8o9 :/fع_IHyÞ$4Q  ʹJYi:1.فo8ЀWDm} f Wn*ѯVevڹub}L[\z:^6 14699ms (09:08:04.886) Jan 26 09:08:04 crc kubenswrapper[4872]: Trace[1784637221]: [14.699201187s] [14.699201187s] END Jan 26 09:08:04 crc kubenswrapper[4872]: I0126 09:08:04.886166 4872 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 26 09:08:04 crc kubenswrapper[4872]: I0126 09:08:04.893551 4872 trace.go:236] Trace[588339445]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Jan-2026 09:07:50.862) (total time: 14030ms): Jan 26 09:08:04 crc kubenswrapper[4872]: Trace[588339445]: ---"Objects listed" error: 14030ms (09:08:04.893) Jan 26 09:08:04 crc kubenswrapper[4872]: Trace[588339445]: [14.030873879s] [14.030873879s] END Jan 26 09:08:04 crc kubenswrapper[4872]: I0126 09:08:04.893634 4872 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 26 09:08:04 crc kubenswrapper[4872]: I0126 09:08:04.894151 4872 trace.go:236] Trace[1132465998]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Jan-2026 09:07:51.089) (total time: 13804ms): Jan 26 09:08:04 crc kubenswrapper[4872]: Trace[1132465998]: ---"Objects listed" error: 13804ms (09:08:04.893) Jan 26 09:08:04 crc kubenswrapper[4872]: Trace[1132465998]: [13.804150951s] [13.804150951s] END Jan 26 09:08:04 crc kubenswrapper[4872]: I0126 09:08:04.894224 4872 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 26 09:08:04 crc kubenswrapper[4872]: I0126 09:08:04.896968 4872 trace.go:236] Trace[2046338767]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Jan-2026 09:07:50.073) (total time: 14823ms): Jan 26 09:08:04 crc kubenswrapper[4872]: Trace[2046338767]: ---"Objects listed" error: 14822ms (09:08:04.896) Jan 26 09:08:04 crc kubenswrapper[4872]: Trace[2046338767]: [14.823145458s] [14.823145458s] END Jan 26 09:08:04 crc kubenswrapper[4872]: I0126 09:08:04.897013 4872 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 26 09:08:04 crc kubenswrapper[4872]: I0126 09:08:04.909378 4872 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Jan 26 09:08:04 crc kubenswrapper[4872]: E0126 09:08:04.911899 4872 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.114298 4872 apiserver.go:52] "Watching apiserver" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.117663 4872 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.118077 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.118541 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.118675 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.118868 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.118917 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.118943 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.119007 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.119010 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.119069 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.119146 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.122206 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.122542 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.122744 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.123492 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.123723 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.123954 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.124186 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.124452 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.125725 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.129101 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 10:34:00.175810811 +0000 UTC Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.130099 4872 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36584->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.130148 4872 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36578->192.168.126.11:17697: read: connection reset by peer" start-of-body= Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.130166 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36584->192.168.126.11:17697: read: connection reset by peer" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.130209 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36578->192.168.126.11:17697: read: connection reset by peer" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.130459 4872 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.130490 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.160695 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.172412 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.183023 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.195042 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.209754 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.218122 4872 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.220410 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.241681 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311438 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311501 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311529 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311555 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311577 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311601 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311622 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311643 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311671 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311689 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311709 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311732 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311757 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311786 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311829 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311846 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311867 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311887 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311924 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311946 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311969 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.311991 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312022 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312051 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312124 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312119 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312152 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312219 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312279 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312322 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312352 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312377 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312399 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312425 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312400 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312449 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312476 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312506 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312534 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312559 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312592 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312618 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312642 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312663 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312675 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312711 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312739 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312763 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312786 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312826 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312848 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312872 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312894 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312919 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312944 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312938 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312970 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313081 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313141 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313170 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313195 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313216 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313235 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313261 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313288 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313307 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313330 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313354 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313374 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313405 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313434 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313461 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313495 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313523 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313548 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313578 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313652 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313674 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313700 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313734 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313765 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313824 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313855 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313881 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313905 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313945 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313970 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313989 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314014 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314043 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314069 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314093 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314113 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314137 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314165 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314188 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314214 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314236 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314255 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314275 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314293 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314320 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314340 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314360 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314381 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314401 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314421 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314440 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314460 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314476 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314493 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314510 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314528 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314545 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314563 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314579 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314598 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314619 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314640 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314705 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314723 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314742 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314775 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314817 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314837 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314858 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314874 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314893 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314913 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314933 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314951 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314969 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314986 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315009 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315032 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315055 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315076 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315105 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315124 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315142 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315161 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315180 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315200 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315220 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315240 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315260 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315287 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315307 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315325 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315409 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315428 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315445 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315470 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315490 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315510 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315529 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315545 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315563 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315579 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315597 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315620 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315636 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315653 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315671 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315691 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315714 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315734 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315755 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315775 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313016 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316098 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316128 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316154 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316176 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316299 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316326 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316347 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316369 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316391 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316412 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316431 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316450 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316468 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316487 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316505 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316526 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316543 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316560 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316580 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316600 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316628 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316655 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316680 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316698 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316717 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316737 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316754 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316773 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316814 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316836 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316856 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316912 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316943 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316967 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316991 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317013 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317034 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317075 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317100 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317131 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317160 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317235 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317258 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317282 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317301 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317394 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317410 4872 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317421 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317435 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317446 4872 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317459 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313031 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.322084 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313089 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313234 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313240 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313269 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313315 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313350 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.312127 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313539 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313669 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313697 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.313854 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314026 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314044 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314104 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314185 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314580 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.314709 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315696 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.315991 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316168 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316246 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316391 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316422 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316653 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316701 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316811 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316874 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.316743 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317094 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.322326 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.322261 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317302 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.322588 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317571 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317813 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317931 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.318144 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.318166 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.318262 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.318309 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.318433 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.318775 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.318444 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.319640 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.320036 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.320154 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.320185 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.320709 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.320739 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.320766 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.321043 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.321060 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.321408 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.321530 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.321577 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.321886 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.322209 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.322242 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.322296 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.317103 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.322564 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.322654 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.322818 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.322844 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.323186 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.323576 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.323685 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.323740 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.323767 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.323885 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.324105 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.324121 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.324202 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.324275 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.324405 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.324432 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.324629 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.324643 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.324650 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.319042 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.325414 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.325462 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.325533 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.325592 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.325688 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:08:05.825655243 +0000 UTC m=+19.134495074 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.325747 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.326667 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.326689 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.327006 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.327657 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.328458 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.327091 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.327139 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.327205 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.328908 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.328935 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.329287 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.329395 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.329426 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.329581 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.329863 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.329915 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.329956 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.330123 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.330273 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.330386 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.330408 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.330396 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.330635 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.330708 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:05.830681103 +0000 UTC m=+19.139521074 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.330965 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.331056 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.331249 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.331748 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.331841 4872 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.331888 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.331899 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.332101 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.332216 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.332306 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.332367 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.332376 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.332410 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.332469 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.332665 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.332725 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.332764 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.332902 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.332917 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:05.832881575 +0000 UTC m=+19.141721566 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.333142 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.333277 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.333540 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.333897 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.334025 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.334154 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.334224 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.334701 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.334859 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.334884 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.335238 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.335319 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.335339 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.335457 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.335934 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.336352 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.336374 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.336401 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.336498 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.336311 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.336775 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.337383 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.337495 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.339974 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.340878 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.340966 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.341047 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.342198 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.344671 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.348419 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.349249 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.349934 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.350840 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.351246 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.351274 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.352669 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.352701 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.352719 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.352719 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.352834 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:05.852777617 +0000 UTC m=+19.161617438 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.353614 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.353962 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.354325 4872 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668" exitCode=255 Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.354399 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668"} Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.354513 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.354943 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.354977 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.354992 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.355051 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:05.855027329 +0000 UTC m=+19.163867130 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.356270 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.358131 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.359978 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.364543 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.365481 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.365836 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.365969 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.366547 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.366729 4872 scope.go:117] "RemoveContainer" containerID="ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.366884 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.367131 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.367365 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.367558 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.367776 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.368000 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.368723 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.368764 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.368849 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.369221 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.369248 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.369202 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.369171 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.369991 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.370435 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.370922 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.371379 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.371487 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.379127 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.379593 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.379943 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.380228 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.380254 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.383612 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.384821 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.391270 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.391399 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.406914 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.417011 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420031 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420152 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420206 4872 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420220 4872 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420231 4872 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420242 4872 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420252 4872 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420265 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420277 4872 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420289 4872 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420299 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420310 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420320 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420329 4872 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420339 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420350 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420362 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420373 4872 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420383 4872 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420392 4872 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420402 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420412 4872 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420421 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420433 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420443 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420452 4872 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420462 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420472 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420482 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420491 4872 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420500 4872 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420509 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420520 4872 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420530 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420540 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420553 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420563 4872 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420573 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420582 4872 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420592 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420602 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420613 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420622 4872 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420632 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420642 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420652 4872 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420662 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420674 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420684 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420694 4872 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420709 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420719 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420729 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420741 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420753 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420764 4872 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420776 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420788 4872 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420813 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420822 4872 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420831 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420841 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420852 4872 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420861 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420871 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420880 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420889 4872 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420899 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420910 4872 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420919 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420929 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420938 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420948 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420958 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420967 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420976 4872 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420986 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.420998 4872 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421007 4872 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421016 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421028 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421038 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421048 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421058 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421068 4872 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421077 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421087 4872 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421097 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421106 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421116 4872 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421125 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421135 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421148 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421158 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421171 4872 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421181 4872 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421190 4872 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421200 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421209 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421218 4872 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421228 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421239 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421248 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421258 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421268 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421278 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421288 4872 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421297 4872 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421307 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421316 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421326 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421337 4872 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421347 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421358 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421368 4872 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421384 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421396 4872 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421411 4872 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421432 4872 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421445 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421458 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421472 4872 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421484 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421495 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421505 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421516 4872 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421526 4872 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421538 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421550 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421562 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421572 4872 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421581 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421591 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421602 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421611 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421620 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421629 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421638 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421647 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421656 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421666 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421675 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421693 4872 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421705 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421716 4872 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421725 4872 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421735 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421745 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421756 4872 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421767 4872 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421779 4872 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421792 4872 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421821 4872 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421831 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421841 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421850 4872 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421860 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421869 4872 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421878 4872 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421888 4872 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421898 4872 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421908 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421919 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421929 4872 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421939 4872 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421949 4872 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421958 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421969 4872 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421978 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421988 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.421998 4872 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422009 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422020 4872 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422033 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422049 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422061 4872 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422074 4872 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422086 4872 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422099 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422113 4872 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422126 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422138 4872 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422151 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422163 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422175 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422189 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422202 4872 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422215 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422227 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422286 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422486 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.422705 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.427139 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.439903 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.449628 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.450479 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.461438 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Jan 26 09:08:05 crc kubenswrapper[4872]: W0126 09:08:05.492718 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-fafc40882e38acde4419f93dccacf813678b3298b57a25b93725025241eb1ef8 WatchSource:0}: Error finding container fafc40882e38acde4419f93dccacf813678b3298b57a25b93725025241eb1ef8: Status 404 returned error can't find the container with id fafc40882e38acde4419f93dccacf813678b3298b57a25b93725025241eb1ef8 Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.522885 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.825905 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.826081 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:08:06.826059956 +0000 UTC m=+20.134899757 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.926814 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.926858 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.926876 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:05 crc kubenswrapper[4872]: I0126 09:08:05.926896 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.927015 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.927032 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.927044 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.927044 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.927104 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:06.927090711 +0000 UTC m=+20.235930512 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.927185 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:06.927149423 +0000 UTC m=+20.235989394 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.927071 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.927241 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:06.927232155 +0000 UTC m=+20.236071956 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.927015 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.927281 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.927301 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:05 crc kubenswrapper[4872]: E0126 09:08:05.927360 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:06.927343998 +0000 UTC m=+20.236184019 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.062645 4872 csr.go:261] certificate signing request csr-dz4q2 is approved, waiting to be issued Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.079387 4872 csr.go:257] certificate signing request csr-dz4q2 is issued Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.129890 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 19:20:59.310540242 +0000 UTC Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.359279 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29"} Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.359363 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"fafc40882e38acde4419f93dccacf813678b3298b57a25b93725025241eb1ef8"} Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.361933 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.365638 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4"} Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.366054 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.372415 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81"} Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.372519 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b"} Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.372534 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5fcf51b6e729ce18066d5201a5880a083ac4e0c75f94d374a31c25ee2d802fee"} Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.374271 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"463ed464e9c099d1d55826cf010394f935e72cce9854d3a6e68315caaed1a202"} Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.377900 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.397911 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.417228 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.431654 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.448087 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.463146 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.479095 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.496246 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.519179 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.546167 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.561125 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.578155 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.597526 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.646660 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:06Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.834207 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.834426 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:08:08.83439648 +0000 UTC m=+22.143236281 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.935346 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.935396 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.935426 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:06 crc kubenswrapper[4872]: I0126 09:08:06.935453 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.935564 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.935571 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.935617 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.935631 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.935636 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:08.93561579 +0000 UTC m=+22.244455591 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.935674 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.935697 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:08.935675952 +0000 UTC m=+22.244515923 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.935825 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:08.935789495 +0000 UTC m=+22.244629296 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.935700 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.935848 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.935862 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:06 crc kubenswrapper[4872]: E0126 09:08:06.935891 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:08.935881757 +0000 UTC m=+22.244721558 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.006955 4872 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.063173 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-gt4gn"] Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.064007 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-6xjb8"] Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.064228 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.064244 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-5sxjr"] Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.064408 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6xjb8" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.065414 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-swcf5"] Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.065921 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.068618 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.073333 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.073441 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.073457 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.073333 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.073730 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.079058 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.079237 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.079065 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.079541 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.079713 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.079770 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.080200 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.081129 4872 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-01-26 09:03:06 +0000 UTC, rotation deadline is 2026-10-19 06:24:07.051386092 +0000 UTC Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.081180 4872 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6381h15m59.970208431s for next certificate rotation Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.081300 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.082068 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.082532 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.096575 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.115861 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.130454 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 06:45:01.88206646 +0000 UTC Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.132774 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.136661 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn756\" (UniqueName: \"kubernetes.io/projected/32b89c0c-4fda-4490-afe4-bf441dd3d337-kube-api-access-pn756\") pod \"node-resolver-6xjb8\" (UID: \"32b89c0c-4fda-4490-afe4-bf441dd3d337\") " pod="openshift-dns/node-resolver-6xjb8" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.136898 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fca44d96-a000-4bf2-8283-a937b0192880-mcd-auth-proxy-config\") pod \"machine-config-daemon-gt4gn\" (UID: \"fca44d96-a000-4bf2-8283-a937b0192880\") " pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.137024 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/732e890d-6c77-4ffa-b74b-1f129c6a96ab-cnibin\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.137168 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-system-cni-dir\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.137299 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-multus-socket-dir-parent\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.137401 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-etc-kubernetes\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.137517 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-os-release\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.137632 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-run-netns\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.137752 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fca44d96-a000-4bf2-8283-a937b0192880-proxy-tls\") pod \"machine-config-daemon-gt4gn\" (UID: \"fca44d96-a000-4bf2-8283-a937b0192880\") " pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.137891 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/732e890d-6c77-4ffa-b74b-1f129c6a96ab-system-cni-dir\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.138005 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbs94\" (UniqueName: \"kubernetes.io/projected/3579107d-a172-4905-a491-6dd2b5a304a5-kube-api-access-gbs94\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.138113 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkwmd\" (UniqueName: \"kubernetes.io/projected/fca44d96-a000-4bf2-8283-a937b0192880-kube-api-access-gkwmd\") pod \"machine-config-daemon-gt4gn\" (UID: \"fca44d96-a000-4bf2-8283-a937b0192880\") " pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.138234 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldz52\" (UniqueName: \"kubernetes.io/projected/732e890d-6c77-4ffa-b74b-1f129c6a96ab-kube-api-access-ldz52\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.138367 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-hostroot\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.138474 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/732e890d-6c77-4ffa-b74b-1f129c6a96ab-os-release\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.138607 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/732e890d-6c77-4ffa-b74b-1f129c6a96ab-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.138699 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-multus-cni-dir\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.138721 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3579107d-a172-4905-a491-6dd2b5a304a5-cni-binary-copy\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.138751 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/732e890d-6c77-4ffa-b74b-1f129c6a96ab-cni-binary-copy\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.138779 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/32b89c0c-4fda-4490-afe4-bf441dd3d337-hosts-file\") pod \"node-resolver-6xjb8\" (UID: \"32b89c0c-4fda-4490-afe4-bf441dd3d337\") " pod="openshift-dns/node-resolver-6xjb8" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.138835 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-var-lib-cni-bin\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.138869 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-multus-conf-dir\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.139024 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-run-k8s-cni-cncf-io\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.139111 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3579107d-a172-4905-a491-6dd2b5a304a5-multus-daemon-config\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.139131 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-cnibin\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.139145 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-var-lib-cni-multus\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.139163 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-var-lib-kubelet\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.139179 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-run-multus-certs\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.139197 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fca44d96-a000-4bf2-8283-a937b0192880-rootfs\") pod \"machine-config-daemon-gt4gn\" (UID: \"fca44d96-a000-4bf2-8283-a937b0192880\") " pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.139214 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/732e890d-6c77-4ffa-b74b-1f129c6a96ab-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.145935 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.162652 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.181597 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.183866 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.183926 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.183937 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:07 crc kubenswrapper[4872]: E0126 09:08:07.184044 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:07 crc kubenswrapper[4872]: E0126 09:08:07.184199 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:07 crc kubenswrapper[4872]: E0126 09:08:07.184303 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.189094 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.189715 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.190584 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.191209 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.191786 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.192429 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.193026 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.193591 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.194235 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.194882 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.195551 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.198548 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.199045 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.199385 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.200052 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.200645 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.201600 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.202154 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.202533 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.203636 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.204263 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.204738 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.205889 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.206315 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.207331 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.207731 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.209032 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.209692 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.210613 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.211190 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.212213 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.212678 4872 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.212779 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.214597 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.215474 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.215942 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.218357 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.218447 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.219532 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.220176 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.221177 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.221932 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.222394 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.223064 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.223730 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.224416 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.224903 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.225459 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.226054 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.226781 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.228575 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.229212 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.230705 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.231776 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.232688 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.233481 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.239795 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3579107d-a172-4905-a491-6dd2b5a304a5-multus-daemon-config\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.239871 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-cnibin\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.239903 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-var-lib-cni-multus\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.239928 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-var-lib-kubelet\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.239955 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-run-multus-certs\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.239989 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/732e890d-6c77-4ffa-b74b-1f129c6a96ab-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240028 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fca44d96-a000-4bf2-8283-a937b0192880-rootfs\") pod \"machine-config-daemon-gt4gn\" (UID: \"fca44d96-a000-4bf2-8283-a937b0192880\") " pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240062 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn756\" (UniqueName: \"kubernetes.io/projected/32b89c0c-4fda-4490-afe4-bf441dd3d337-kube-api-access-pn756\") pod \"node-resolver-6xjb8\" (UID: \"32b89c0c-4fda-4490-afe4-bf441dd3d337\") " pod="openshift-dns/node-resolver-6xjb8" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240106 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fca44d96-a000-4bf2-8283-a937b0192880-mcd-auth-proxy-config\") pod \"machine-config-daemon-gt4gn\" (UID: \"fca44d96-a000-4bf2-8283-a937b0192880\") " pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240143 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-system-cni-dir\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240155 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-var-lib-kubelet\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240176 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-multus-socket-dir-parent\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240209 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-cnibin\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240216 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-etc-kubernetes\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240235 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-etc-kubernetes\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240256 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-multus-socket-dir-parent\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240266 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-var-lib-cni-multus\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240258 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/732e890d-6c77-4ffa-b74b-1f129c6a96ab-cnibin\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240281 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/732e890d-6c77-4ffa-b74b-1f129c6a96ab-cnibin\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240306 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fca44d96-a000-4bf2-8283-a937b0192880-rootfs\") pod \"machine-config-daemon-gt4gn\" (UID: \"fca44d96-a000-4bf2-8283-a937b0192880\") " pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240349 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-system-cni-dir\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240404 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-run-multus-certs\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.240685 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/732e890d-6c77-4ffa-b74b-1f129c6a96ab-tuning-conf-dir\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.241115 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-os-release\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.241160 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-run-netns\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.241194 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fca44d96-a000-4bf2-8283-a937b0192880-proxy-tls\") pod \"machine-config-daemon-gt4gn\" (UID: \"fca44d96-a000-4bf2-8283-a937b0192880\") " pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.241215 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/732e890d-6c77-4ffa-b74b-1f129c6a96ab-system-cni-dir\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.241217 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3579107d-a172-4905-a491-6dd2b5a304a5-multus-daemon-config\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.241248 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-hostroot\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.241269 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbs94\" (UniqueName: \"kubernetes.io/projected/3579107d-a172-4905-a491-6dd2b5a304a5-kube-api-access-gbs94\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.241291 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkwmd\" (UniqueName: \"kubernetes.io/projected/fca44d96-a000-4bf2-8283-a937b0192880-kube-api-access-gkwmd\") pod \"machine-config-daemon-gt4gn\" (UID: \"fca44d96-a000-4bf2-8283-a937b0192880\") " pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.241314 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldz52\" (UniqueName: \"kubernetes.io/projected/732e890d-6c77-4ffa-b74b-1f129c6a96ab-kube-api-access-ldz52\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.241341 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fca44d96-a000-4bf2-8283-a937b0192880-mcd-auth-proxy-config\") pod \"machine-config-daemon-gt4gn\" (UID: \"fca44d96-a000-4bf2-8283-a937b0192880\") " pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.242937 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-run-netns\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243019 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-multus-cni-dir\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243101 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-hostroot\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243121 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-multus-cni-dir\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243124 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/732e890d-6c77-4ffa-b74b-1f129c6a96ab-system-cni-dir\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243155 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3579107d-a172-4905-a491-6dd2b5a304a5-cni-binary-copy\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243187 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/732e890d-6c77-4ffa-b74b-1f129c6a96ab-os-release\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243214 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/732e890d-6c77-4ffa-b74b-1f129c6a96ab-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243261 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/732e890d-6c77-4ffa-b74b-1f129c6a96ab-cni-binary-copy\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243328 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/32b89c0c-4fda-4490-afe4-bf441dd3d337-hosts-file\") pod \"node-resolver-6xjb8\" (UID: \"32b89c0c-4fda-4490-afe4-bf441dd3d337\") " pod="openshift-dns/node-resolver-6xjb8" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243387 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/732e890d-6c77-4ffa-b74b-1f129c6a96ab-os-release\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243582 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/32b89c0c-4fda-4490-afe4-bf441dd3d337-hosts-file\") pod \"node-resolver-6xjb8\" (UID: \"32b89c0c-4fda-4490-afe4-bf441dd3d337\") " pod="openshift-dns/node-resolver-6xjb8" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243640 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-multus-conf-dir\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243665 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-var-lib-cni-bin\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243692 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-run-k8s-cni-cncf-io\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243787 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-run-k8s-cni-cncf-io\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243812 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-multus-conf-dir\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.243846 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-host-var-lib-cni-bin\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.244205 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3579107d-a172-4905-a491-6dd2b5a304a5-cni-binary-copy\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.244266 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/732e890d-6c77-4ffa-b74b-1f129c6a96ab-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.245856 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3579107d-a172-4905-a491-6dd2b5a304a5-os-release\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.247121 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/732e890d-6c77-4ffa-b74b-1f129c6a96ab-cni-binary-copy\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.247506 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.249393 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fca44d96-a000-4bf2-8283-a937b0192880-proxy-tls\") pod \"machine-config-daemon-gt4gn\" (UID: \"fca44d96-a000-4bf2-8283-a937b0192880\") " pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.262333 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn756\" (UniqueName: \"kubernetes.io/projected/32b89c0c-4fda-4490-afe4-bf441dd3d337-kube-api-access-pn756\") pod \"node-resolver-6xjb8\" (UID: \"32b89c0c-4fda-4490-afe4-bf441dd3d337\") " pod="openshift-dns/node-resolver-6xjb8" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.264017 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldz52\" (UniqueName: \"kubernetes.io/projected/732e890d-6c77-4ffa-b74b-1f129c6a96ab-kube-api-access-ldz52\") pod \"multus-additional-cni-plugins-5sxjr\" (UID: \"732e890d-6c77-4ffa-b74b-1f129c6a96ab\") " pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.265194 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkwmd\" (UniqueName: \"kubernetes.io/projected/fca44d96-a000-4bf2-8283-a937b0192880-kube-api-access-gkwmd\") pod \"machine-config-daemon-gt4gn\" (UID: \"fca44d96-a000-4bf2-8283-a937b0192880\") " pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.271194 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbs94\" (UniqueName: \"kubernetes.io/projected/3579107d-a172-4905-a491-6dd2b5a304a5-kube-api-access-gbs94\") pod \"multus-swcf5\" (UID: \"3579107d-a172-4905-a491-6dd2b5a304a5\") " pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.277255 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.298376 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.316890 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.330553 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.346501 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.361888 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.372972 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.382592 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.391352 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.394504 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-6xjb8" Jan 26 09:08:07 crc kubenswrapper[4872]: W0126 09:08:07.397367 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfca44d96_a000_4bf2_8283_a937b0192880.slice/crio-f472628d98a56b3c24a6dbcd8a3e8e97209a8c9b64cc0c702930ca4484dce583 WatchSource:0}: Error finding container f472628d98a56b3c24a6dbcd8a3e8e97209a8c9b64cc0c702930ca4484dce583: Status 404 returned error can't find the container with id f472628d98a56b3c24a6dbcd8a3e8e97209a8c9b64cc0c702930ca4484dce583 Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.403722 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" Jan 26 09:08:07 crc kubenswrapper[4872]: W0126 09:08:07.407569 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32b89c0c_4fda_4490_afe4_bf441dd3d337.slice/crio-a2cfebd0bc6dca1c317f849410ebfa2daee93a4df6062dc352264a37a02b4549 WatchSource:0}: Error finding container a2cfebd0bc6dca1c317f849410ebfa2daee93a4df6062dc352264a37a02b4549: Status 404 returned error can't find the container with id a2cfebd0bc6dca1c317f849410ebfa2daee93a4df6062dc352264a37a02b4549 Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.411601 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-swcf5" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.413207 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.429963 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.447457 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.459524 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6p5bj"] Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.460719 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.463438 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.463541 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.466171 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.466311 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.466367 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.466514 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.466679 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.479376 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.495101 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.515867 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546169 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546219 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-var-lib-openvswitch\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546242 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-env-overrides\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546316 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-ovn\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546344 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-systemd-units\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546385 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-node-log\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546402 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-cni-netd\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546420 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtxcp\" (UniqueName: \"kubernetes.io/projected/e2577331-1886-4abb-896b-8d8a112e40d8-kube-api-access-qtxcp\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546435 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-ovnkube-config\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546468 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e2577331-1886-4abb-896b-8d8a112e40d8-ovn-node-metrics-cert\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546484 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-systemd\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546520 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-run-netns\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546575 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-cni-bin\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546597 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-log-socket\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546642 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-etc-openvswitch\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546670 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-slash\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546719 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-openvswitch\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546768 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-run-ovn-kubernetes\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546837 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-kubelet\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.546887 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-ovnkube-script-lib\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.549004 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.584211 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.612932 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.629041 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.642845 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647367 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-kubelet\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647407 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-run-ovn-kubernetes\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647436 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-ovnkube-script-lib\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647462 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647482 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-var-lib-openvswitch\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647497 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-env-overrides\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647517 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-ovn\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647537 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-systemd-units\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647535 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-run-ovn-kubernetes\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647568 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647553 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-node-log\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647605 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-ovn\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647604 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-node-log\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647571 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-var-lib-openvswitch\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647641 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-cni-netd\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647645 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-systemd-units\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647685 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtxcp\" (UniqueName: \"kubernetes.io/projected/e2577331-1886-4abb-896b-8d8a112e40d8-kube-api-access-qtxcp\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647681 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-cni-netd\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647723 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-ovnkube-config\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647744 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e2577331-1886-4abb-896b-8d8a112e40d8-ovn-node-metrics-cert\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647772 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-systemd\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647808 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-run-netns\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647826 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-cni-bin\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647846 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-log-socket\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647869 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-etc-openvswitch\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647885 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-slash\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647907 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-openvswitch\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.647990 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-openvswitch\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.648032 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-systemd\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.648059 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-run-netns\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.648084 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-etc-openvswitch\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.648090 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-log-socket\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.648108 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-slash\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.648133 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-cni-bin\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.648149 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-kubelet\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.648261 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-env-overrides\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.648377 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-ovnkube-config\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.648844 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-ovnkube-script-lib\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.653034 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e2577331-1886-4abb-896b-8d8a112e40d8-ovn-node-metrics-cert\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.654639 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.667747 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtxcp\" (UniqueName: \"kubernetes.io/projected/e2577331-1886-4abb-896b-8d8a112e40d8-kube-api-access-qtxcp\") pod \"ovnkube-node-6p5bj\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.672383 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.696531 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.712690 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.716508 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.718138 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.725866 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.735594 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.753949 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.767641 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.772173 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.772958 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: W0126 09:08:07.784247 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2577331_1886_4abb_896b_8d8a112e40d8.slice/crio-457a75a10afa7862edbcafd63b2fcc0a7ba8b01cb2bd81e8340eea6fba0e71af WatchSource:0}: Error finding container 457a75a10afa7862edbcafd63b2fcc0a7ba8b01cb2bd81e8340eea6fba0e71af: Status 404 returned error can't find the container with id 457a75a10afa7862edbcafd63b2fcc0a7ba8b01cb2bd81e8340eea6fba0e71af Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.786083 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.794031 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.812690 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.831909 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.840220 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.860644 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.879256 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.897354 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.916117 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.929464 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.954000 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.967897 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.982529 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:07 crc kubenswrapper[4872]: I0126 09:08:07.995743 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.006469 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.020459 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.033624 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.048391 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.065294 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.079031 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.091878 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.104727 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.112039 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.114266 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.114310 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.114323 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.114494 4872 kubelet_node_status.go:76] "Attempting to register node" node="crc" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.121506 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.122655 4872 kubelet_node_status.go:115] "Node was previously registered" node="crc" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.123014 4872 kubelet_node_status.go:79] "Successfully registered node" node="crc" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.124189 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.124241 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.124258 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.124280 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.124295 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.130730 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 20:52:22.057211972 +0000 UTC Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.142463 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.146405 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.146457 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.146469 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.146490 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.146503 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.160392 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.164425 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.164501 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.164513 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.164536 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.164552 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.178760 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.183240 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.183298 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.183309 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.183323 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.183333 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.196492 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.199945 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.200008 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.200027 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.200050 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.200067 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.215521 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.215650 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.217850 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.217894 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.217904 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.217921 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.217932 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.320748 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.320822 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.320838 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.320860 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.320874 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.382993 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.383073 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"457a75a10afa7862edbcafd63b2fcc0a7ba8b01cb2bd81e8340eea6fba0e71af"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.384644 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" event={"ID":"732e890d-6c77-4ffa-b74b-1f129c6a96ab","Type":"ContainerStarted","Data":"8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.384705 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" event={"ID":"732e890d-6c77-4ffa-b74b-1f129c6a96ab","Type":"ContainerStarted","Data":"06dd3832e8b7b07fb533df4671a59a76fa30708f306e15ce811c1a310eebcfdc"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.387133 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerStarted","Data":"ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.387238 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerStarted","Data":"28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.387260 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerStarted","Data":"f472628d98a56b3c24a6dbcd8a3e8e97209a8c9b64cc0c702930ca4484dce583"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.388999 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-swcf5" event={"ID":"3579107d-a172-4905-a491-6dd2b5a304a5","Type":"ContainerStarted","Data":"27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.389040 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-swcf5" event={"ID":"3579107d-a172-4905-a491-6dd2b5a304a5","Type":"ContainerStarted","Data":"c0f5f82fa9657c4ef95693874f6c028b33e3b3ad0650511ad8a520335758ca10"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.390258 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6xjb8" event={"ID":"32b89c0c-4fda-4490-afe4-bf441dd3d337","Type":"ContainerStarted","Data":"8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.390350 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-6xjb8" event={"ID":"32b89c0c-4fda-4490-afe4-bf441dd3d337","Type":"ContainerStarted","Data":"a2cfebd0bc6dca1c317f849410ebfa2daee93a4df6062dc352264a37a02b4549"} Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.413202 4872 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.423750 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.423829 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.423850 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.423876 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.423892 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.425908 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.465247 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.483197 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.495419 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.506793 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.521101 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.526685 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.526729 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.526741 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.526762 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.526774 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.542015 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.559983 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.577437 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.594427 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.606564 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.617829 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.629183 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.629224 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.629236 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.629256 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.629270 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.632318 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.652108 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.677709 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.694099 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.712793 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.728498 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.732488 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.732663 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.732723 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.732824 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.732908 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.743783 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.761542 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.797151 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.835915 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.835977 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.835990 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.836014 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.836035 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.842603 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.860353 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.860536 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:08:12.86050734 +0000 UTC m=+26.169347141 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.876734 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.916487 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.939660 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.939720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.939735 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.939759 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.939776 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:08Z","lastTransitionTime":"2026-01-26T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.961541 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.961596 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.961628 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.961663 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.961761 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.961843 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:12.961824733 +0000 UTC m=+26.270664534 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.961932 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.962033 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:12.962016398 +0000 UTC m=+26.270856199 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.962044 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.962096 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.962112 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.962044 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.962230 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.962246 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.962199 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:12.962170333 +0000 UTC m=+26.271010314 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:08 crc kubenswrapper[4872]: E0126 09:08:08.962308 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:12.962290726 +0000 UTC m=+26.271130727 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.964315 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:08 crc kubenswrapper[4872]: I0126 09:08:08.996971 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:08Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.035441 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.045980 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.046042 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.046070 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.046095 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.046106 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:09Z","lastTransitionTime":"2026-01-26T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.079265 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.131751 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 21:39:49.234136115 +0000 UTC Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.149582 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.149647 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.149660 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.149686 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.149702 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:09Z","lastTransitionTime":"2026-01-26T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.185636 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:09 crc kubenswrapper[4872]: E0126 09:08:09.185775 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.186152 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:09 crc kubenswrapper[4872]: E0126 09:08:09.186210 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.186249 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:09 crc kubenswrapper[4872]: E0126 09:08:09.186303 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.253387 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.253430 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.253440 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.253459 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.253469 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:09Z","lastTransitionTime":"2026-01-26T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.357513 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.357595 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.357685 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.357716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.357735 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:09Z","lastTransitionTime":"2026-01-26T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.396863 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.398346 4872 generic.go:334] "Generic (PLEG): container finished" podID="e2577331-1886-4abb-896b-8d8a112e40d8" containerID="5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c" exitCode=0 Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.398440 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.399755 4872 generic.go:334] "Generic (PLEG): container finished" podID="732e890d-6c77-4ffa-b74b-1f129c6a96ab" containerID="8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e" exitCode=0 Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.399876 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" event={"ID":"732e890d-6c77-4ffa-b74b-1f129c6a96ab","Type":"ContainerDied","Data":"8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.421478 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.446071 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.459075 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.460705 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.460755 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.460771 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.460818 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.460838 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:09Z","lastTransitionTime":"2026-01-26T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.472666 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.485696 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.501149 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.518098 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.533314 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.551120 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.564490 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.564538 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.564548 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.564574 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.564587 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:09Z","lastTransitionTime":"2026-01-26T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.578404 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.596126 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.612242 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.626035 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.667358 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.667404 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.667415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.667432 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.667446 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:09Z","lastTransitionTime":"2026-01-26T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.692212 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.725402 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.742397 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.760164 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.770023 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.770071 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.770084 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.770102 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.770116 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:09Z","lastTransitionTime":"2026-01-26T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.796908 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.836052 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.873134 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.873306 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.873368 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.873434 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.873489 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:09Z","lastTransitionTime":"2026-01-26T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.875777 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.922101 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.959638 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.976449 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.976497 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.976507 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.976527 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.976540 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:09Z","lastTransitionTime":"2026-01-26T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:09 crc kubenswrapper[4872]: I0126 09:08:09.995335 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:09Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.035786 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.079890 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.079947 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.079961 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.079991 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.080005 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:10Z","lastTransitionTime":"2026-01-26T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.081793 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.119349 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.132052 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 14:35:38.288059275 +0000 UTC Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.132486 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-d66vd"] Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.132994 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-d66vd" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.163858 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.166678 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.174719 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5461db0d-f9c7-4215-b259-9e10e2d318df-serviceca\") pod \"node-ca-d66vd\" (UID: \"5461db0d-f9c7-4215-b259-9e10e2d318df\") " pod="openshift-image-registry/node-ca-d66vd" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.174765 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5461db0d-f9c7-4215-b259-9e10e2d318df-host\") pod \"node-ca-d66vd\" (UID: \"5461db0d-f9c7-4215-b259-9e10e2d318df\") " pod="openshift-image-registry/node-ca-d66vd" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.174831 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2c7c\" (UniqueName: \"kubernetes.io/projected/5461db0d-f9c7-4215-b259-9e10e2d318df-kube-api-access-x2c7c\") pod \"node-ca-d66vd\" (UID: \"5461db0d-f9c7-4215-b259-9e10e2d318df\") " pod="openshift-image-registry/node-ca-d66vd" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.182700 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.182743 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.182753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.182776 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.182787 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:10Z","lastTransitionTime":"2026-01-26T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.187267 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.206576 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.226538 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.278363 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.317120 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.356322 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.362416 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5461db0d-f9c7-4215-b259-9e10e2d318df-host\") pod \"node-ca-d66vd\" (UID: \"5461db0d-f9c7-4215-b259-9e10e2d318df\") " pod="openshift-image-registry/node-ca-d66vd" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.362532 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2c7c\" (UniqueName: \"kubernetes.io/projected/5461db0d-f9c7-4215-b259-9e10e2d318df-kube-api-access-x2c7c\") pod \"node-ca-d66vd\" (UID: \"5461db0d-f9c7-4215-b259-9e10e2d318df\") " pod="openshift-image-registry/node-ca-d66vd" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.362569 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5461db0d-f9c7-4215-b259-9e10e2d318df-serviceca\") pod \"node-ca-d66vd\" (UID: \"5461db0d-f9c7-4215-b259-9e10e2d318df\") " pod="openshift-image-registry/node-ca-d66vd" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.362693 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5461db0d-f9c7-4215-b259-9e10e2d318df-host\") pod \"node-ca-d66vd\" (UID: \"5461db0d-f9c7-4215-b259-9e10e2d318df\") " pod="openshift-image-registry/node-ca-d66vd" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.364046 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5461db0d-f9c7-4215-b259-9e10e2d318df-serviceca\") pod \"node-ca-d66vd\" (UID: \"5461db0d-f9c7-4215-b259-9e10e2d318df\") " pod="openshift-image-registry/node-ca-d66vd" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.364720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.364767 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.364789 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.364854 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.364872 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:10Z","lastTransitionTime":"2026-01-26T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.406134 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2c7c\" (UniqueName: \"kubernetes.io/projected/5461db0d-f9c7-4215-b259-9e10e2d318df-kube-api-access-x2c7c\") pod \"node-ca-d66vd\" (UID: \"5461db0d-f9c7-4215-b259-9e10e2d318df\") " pod="openshift-image-registry/node-ca-d66vd" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.409836 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.409892 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.409905 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.409915 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.409924 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.409933 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.411574 4872 generic.go:334] "Generic (PLEG): container finished" podID="732e890d-6c77-4ffa-b74b-1f129c6a96ab" containerID="f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30" exitCode=0 Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.411653 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" event={"ID":"732e890d-6c77-4ffa-b74b-1f129c6a96ab","Type":"ContainerDied","Data":"f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.420431 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.454368 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-d66vd" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.456647 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.468077 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.468129 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.468140 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.468164 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.468177 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:10Z","lastTransitionTime":"2026-01-26T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:10 crc kubenswrapper[4872]: W0126 09:08:10.473744 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5461db0d_f9c7_4215_b259_9e10e2d318df.slice/crio-eba00d9c85711de093294c81cc474318c7bf9f76449a5dfbf073b2b267612ce4 WatchSource:0}: Error finding container eba00d9c85711de093294c81cc474318c7bf9f76449a5dfbf073b2b267612ce4: Status 404 returned error can't find the container with id eba00d9c85711de093294c81cc474318c7bf9f76449a5dfbf073b2b267612ce4 Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.494565 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.534106 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.572725 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.572772 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.572784 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.572830 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.572843 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:10Z","lastTransitionTime":"2026-01-26T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.580615 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.614629 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.658421 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.675395 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.675434 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.675446 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.675464 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.675476 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:10Z","lastTransitionTime":"2026-01-26T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.697091 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.734565 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.777407 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.779473 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.779503 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.779511 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.779527 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.779538 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:10Z","lastTransitionTime":"2026-01-26T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.813702 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.874635 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.881492 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.881546 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.881561 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.881582 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.881597 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:10Z","lastTransitionTime":"2026-01-26T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.895244 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.936571 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.984664 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.984739 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.984752 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.984770 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.984779 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:10Z","lastTransitionTime":"2026-01-26T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:10 crc kubenswrapper[4872]: I0126 09:08:10.985895 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.020671 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.057224 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.087335 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.087380 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.087390 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.087407 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.087418 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:11Z","lastTransitionTime":"2026-01-26T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.098161 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.132895 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 01:13:27.697066977 +0000 UTC Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.134339 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.177949 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.183182 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:11 crc kubenswrapper[4872]: E0126 09:08:11.183391 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.183198 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:11 crc kubenswrapper[4872]: E0126 09:08:11.183511 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.183190 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:11 crc kubenswrapper[4872]: E0126 09:08:11.183608 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.189383 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.189416 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.189423 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.189437 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.189447 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:11Z","lastTransitionTime":"2026-01-26T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.216039 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.256907 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.291841 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.291886 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.291899 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.291916 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.291935 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:11Z","lastTransitionTime":"2026-01-26T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.304014 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.334768 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.377073 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.394587 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.394650 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.394673 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.394710 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.394733 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:11Z","lastTransitionTime":"2026-01-26T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.417550 4872 generic.go:334] "Generic (PLEG): container finished" podID="732e890d-6c77-4ffa-b74b-1f129c6a96ab" containerID="5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c" exitCode=0 Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.417640 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" event={"ID":"732e890d-6c77-4ffa-b74b-1f129c6a96ab","Type":"ContainerDied","Data":"5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c"} Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.420328 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-d66vd" event={"ID":"5461db0d-f9c7-4215-b259-9e10e2d318df","Type":"ContainerStarted","Data":"8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec"} Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.420358 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-d66vd" event={"ID":"5461db0d-f9c7-4215-b259-9e10e2d318df","Type":"ContainerStarted","Data":"eba00d9c85711de093294c81cc474318c7bf9f76449a5dfbf073b2b267612ce4"} Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.426277 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.468917 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.497604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.497652 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.497661 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.497678 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.497688 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:11Z","lastTransitionTime":"2026-01-26T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.503166 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.542957 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.574117 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.600136 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.600173 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.600181 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.600196 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.600206 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:11Z","lastTransitionTime":"2026-01-26T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.615934 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.657002 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.696410 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.703250 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.703306 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.703326 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.703350 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.703367 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:11Z","lastTransitionTime":"2026-01-26T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.739562 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.777363 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.806285 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.806328 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.806340 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.806356 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.806369 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:11Z","lastTransitionTime":"2026-01-26T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.824326 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.856630 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.899821 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.909399 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.909429 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.909440 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.909453 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.909462 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:11Z","lastTransitionTime":"2026-01-26T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.934918 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:11 crc kubenswrapper[4872]: I0126 09:08:11.975832 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:11Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.012266 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.012311 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.012320 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.012339 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.012350 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:12Z","lastTransitionTime":"2026-01-26T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.020173 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.056010 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.094401 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.114622 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.114654 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.114662 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.114676 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.114686 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:12Z","lastTransitionTime":"2026-01-26T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.133471 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 11:27:14.754461441 +0000 UTC Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.217027 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.217078 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.217089 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.217115 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.217126 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:12Z","lastTransitionTime":"2026-01-26T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.319616 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.319664 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.319677 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.319694 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.319706 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:12Z","lastTransitionTime":"2026-01-26T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.421767 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.421818 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.421827 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.421839 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.421848 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:12Z","lastTransitionTime":"2026-01-26T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.424207 4872 generic.go:334] "Generic (PLEG): container finished" podID="732e890d-6c77-4ffa-b74b-1f129c6a96ab" containerID="648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a" exitCode=0 Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.424260 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" event={"ID":"732e890d-6c77-4ffa-b74b-1f129c6a96ab","Type":"ContainerDied","Data":"648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a"} Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.430589 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665"} Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.441115 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.456254 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.473749 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.487517 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.501217 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.512289 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.524050 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.524087 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.524099 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.524117 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.524131 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:12Z","lastTransitionTime":"2026-01-26T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.533203 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.545930 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.561443 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.577316 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.591986 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.606550 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.619420 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.626464 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.626506 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.626522 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.626543 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.626556 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:12Z","lastTransitionTime":"2026-01-26T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.662944 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.696935 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.730833 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.731267 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.731275 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.731294 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.731306 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:12Z","lastTransitionTime":"2026-01-26T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.834749 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.834878 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.834905 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.834936 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.834961 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:12Z","lastTransitionTime":"2026-01-26T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.884031 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.884253 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:08:20.884222829 +0000 UTC m=+34.193062680 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.937978 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.938025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.938035 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.938053 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.938064 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:12Z","lastTransitionTime":"2026-01-26T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.985410 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.985461 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.985492 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:12 crc kubenswrapper[4872]: I0126 09:08:12.985515 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.985595 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.985630 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.985648 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.985657 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:20.985640994 +0000 UTC m=+34.294480795 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.985665 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.985746 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:20.985726236 +0000 UTC m=+34.294566067 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.985770 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.985926 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:20.985905191 +0000 UTC m=+34.294745022 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.985961 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.986040 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.986069 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:12 crc kubenswrapper[4872]: E0126 09:08:12.986208 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:20.986168609 +0000 UTC m=+34.295008540 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.041319 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.041395 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.041413 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.041439 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.041468 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:13Z","lastTransitionTime":"2026-01-26T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.134195 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 04:37:47.744175995 +0000 UTC Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.144399 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.144474 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.144495 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.144522 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.144541 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:13Z","lastTransitionTime":"2026-01-26T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.183911 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.183992 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.183937 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:13 crc kubenswrapper[4872]: E0126 09:08:13.184086 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:13 crc kubenswrapper[4872]: E0126 09:08:13.184194 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:13 crc kubenswrapper[4872]: E0126 09:08:13.184256 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.247855 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.247931 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.247952 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.247982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.248002 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:13Z","lastTransitionTime":"2026-01-26T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.351313 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.351388 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.351410 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.351442 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.351461 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:13Z","lastTransitionTime":"2026-01-26T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.438784 4872 generic.go:334] "Generic (PLEG): container finished" podID="732e890d-6c77-4ffa-b74b-1f129c6a96ab" containerID="79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb" exitCode=0 Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.438884 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" event={"ID":"732e890d-6c77-4ffa-b74b-1f129c6a96ab","Type":"ContainerDied","Data":"79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb"} Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.454639 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.454724 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.454752 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.454785 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.454852 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:13Z","lastTransitionTime":"2026-01-26T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.461022 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.475917 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.486854 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.503120 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.524675 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.548078 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.556758 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.556831 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.556845 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.556863 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.556877 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:13Z","lastTransitionTime":"2026-01-26T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.579143 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.595025 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.620886 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.642601 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.660423 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.661208 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.661300 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.661318 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.661342 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.661394 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:13Z","lastTransitionTime":"2026-01-26T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.681596 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.697753 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.717686 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.733607 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:13Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.763757 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.763832 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.763845 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.763862 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.763875 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:13Z","lastTransitionTime":"2026-01-26T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.867685 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.867871 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.867902 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.867937 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.867974 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:13Z","lastTransitionTime":"2026-01-26T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.972625 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.972733 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.972752 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.972776 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:13 crc kubenswrapper[4872]: I0126 09:08:13.972790 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:13Z","lastTransitionTime":"2026-01-26T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.076351 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.076428 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.076451 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.076489 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.076514 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:14Z","lastTransitionTime":"2026-01-26T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.135084 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 12:41:35.721421815 +0000 UTC Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.179984 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.180052 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.180070 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.180096 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.180114 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:14Z","lastTransitionTime":"2026-01-26T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.283904 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.284042 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.284074 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.284173 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.284253 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:14Z","lastTransitionTime":"2026-01-26T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.387238 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.387296 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.387308 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.387326 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.387341 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:14Z","lastTransitionTime":"2026-01-26T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.446774 4872 generic.go:334] "Generic (PLEG): container finished" podID="732e890d-6c77-4ffa-b74b-1f129c6a96ab" containerID="065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac" exitCode=0 Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.446917 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" event={"ID":"732e890d-6c77-4ffa-b74b-1f129c6a96ab","Type":"ContainerDied","Data":"065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac"} Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.474160 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.492179 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.492240 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.492259 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.492290 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.492316 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:14Z","lastTransitionTime":"2026-01-26T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.505859 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.527084 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.567247 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.587199 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.597922 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.597969 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.597979 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.597997 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.598008 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:14Z","lastTransitionTime":"2026-01-26T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.607397 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.633768 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.650003 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.665528 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.686198 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.697500 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.700383 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.700441 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.700452 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.700477 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.700491 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:14Z","lastTransitionTime":"2026-01-26T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.713542 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.726984 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.738893 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.755949 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:14Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.802877 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.802934 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.802947 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.802972 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.802988 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:14Z","lastTransitionTime":"2026-01-26T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.907323 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.907376 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.907387 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.907406 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:14 crc kubenswrapper[4872]: I0126 09:08:14.907421 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:14Z","lastTransitionTime":"2026-01-26T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.010268 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.010641 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.010660 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.010698 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.010714 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:15Z","lastTransitionTime":"2026-01-26T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.116162 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.116223 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.116235 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.116256 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.116269 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:15Z","lastTransitionTime":"2026-01-26T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.135389 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 18:07:35.171241705 +0000 UTC Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.184193 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:15 crc kubenswrapper[4872]: E0126 09:08:15.184463 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.184268 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:15 crc kubenswrapper[4872]: E0126 09:08:15.184622 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.184207 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:15 crc kubenswrapper[4872]: E0126 09:08:15.184736 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.218583 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.218660 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.218678 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.218705 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.218725 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:15Z","lastTransitionTime":"2026-01-26T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.323180 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.323274 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.323299 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.323335 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.323361 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:15Z","lastTransitionTime":"2026-01-26T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.469852 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.469937 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.469963 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.469999 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.470024 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:15Z","lastTransitionTime":"2026-01-26T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.477770 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.478019 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.478073 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.484144 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" event={"ID":"732e890d-6c77-4ffa-b74b-1f129c6a96ab","Type":"ContainerStarted","Data":"a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.495778 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.520323 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.542477 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.561939 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.573764 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.573855 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.573872 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.573891 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.573901 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:15Z","lastTransitionTime":"2026-01-26T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.578838 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.594060 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.602362 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.602426 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.610429 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.632212 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.644010 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.663418 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.675080 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.676612 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.676670 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.676688 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.676714 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.676733 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:15Z","lastTransitionTime":"2026-01-26T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.689382 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.712945 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.730579 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.750236 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.771275 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.779545 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.779619 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.779635 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.779662 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.779679 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:15Z","lastTransitionTime":"2026-01-26T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.794578 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.814662 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.835964 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.861320 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.882682 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.882744 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.882761 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.882787 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.882837 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:15Z","lastTransitionTime":"2026-01-26T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.904699 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.922911 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.938033 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.958578 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.973531 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.985753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.985850 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.985870 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.985895 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.985911 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:15Z","lastTransitionTime":"2026-01-26T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:15 crc kubenswrapper[4872]: I0126 09:08:15.989663 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:15Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.004791 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.016718 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.043178 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.060637 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.088553 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.088613 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.088637 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.088670 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.088694 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:16Z","lastTransitionTime":"2026-01-26T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.135558 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 10:48:20.739352813 +0000 UTC Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.192438 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.192491 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.192508 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.192530 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.192547 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:16Z","lastTransitionTime":"2026-01-26T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.262283 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.290626 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.295901 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.295969 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.295995 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.296026 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.296050 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:16Z","lastTransitionTime":"2026-01-26T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.327070 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.347199 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.362456 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.378268 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.398207 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.399403 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.399499 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.399522 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.399567 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.399590 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:16Z","lastTransitionTime":"2026-01-26T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.418654 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.433842 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.452700 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.473433 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.487990 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.489551 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.502461 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.502499 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.502512 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.502544 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.502558 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:16Z","lastTransitionTime":"2026-01-26T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.506398 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.524404 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.538072 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.555000 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:16Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.605378 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.605422 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.605432 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.605452 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.605463 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:16Z","lastTransitionTime":"2026-01-26T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.707906 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.707950 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.707959 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.707980 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.707991 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:16Z","lastTransitionTime":"2026-01-26T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.810727 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.810782 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.810793 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.810835 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.810850 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:16Z","lastTransitionTime":"2026-01-26T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.913420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.913479 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.913497 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.913523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:16 crc kubenswrapper[4872]: I0126 09:08:16.913540 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:16Z","lastTransitionTime":"2026-01-26T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.016427 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.016825 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.016968 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.017104 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.017224 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:17Z","lastTransitionTime":"2026-01-26T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.121961 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.122033 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.122051 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.122076 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.122098 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:17Z","lastTransitionTime":"2026-01-26T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.135858 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 23:01:17.432128443 +0000 UTC Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.183353 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.183460 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:17 crc kubenswrapper[4872]: E0126 09:08:17.183561 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.183843 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:17 crc kubenswrapper[4872]: E0126 09:08:17.183944 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:17 crc kubenswrapper[4872]: E0126 09:08:17.184126 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.221771 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.225306 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.225376 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.225401 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.225435 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.225524 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:17Z","lastTransitionTime":"2026-01-26T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.240454 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.265688 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.288214 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.308235 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.326186 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.328218 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.328284 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.328293 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.328307 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.328316 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:17Z","lastTransitionTime":"2026-01-26T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.341242 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: E0126 09:08:17.342109 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2577331_1886_4abb_896b_8d8a112e40d8.slice/crio-3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74.scope\": RecentStats: unable to find data in memory cache]" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.360173 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.373293 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.384328 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.397712 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.416129 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.430203 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.430851 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.430895 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.430906 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.430925 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.430938 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:17Z","lastTransitionTime":"2026-01-26T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.444398 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.461197 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.493884 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovnkube-controller/0.log" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.496928 4872 generic.go:334] "Generic (PLEG): container finished" podID="e2577331-1886-4abb-896b-8d8a112e40d8" containerID="3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74" exitCode=1 Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.496968 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74"} Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.497968 4872 scope.go:117] "RemoveContainer" containerID="3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.520405 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.534152 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.537237 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.537269 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.537290 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.537309 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:17Z","lastTransitionTime":"2026-01-26T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.540787 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.557535 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.579061 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.594554 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.620979 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.636334 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.640347 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.640380 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.640390 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.640407 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.640420 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:17Z","lastTransitionTime":"2026-01-26T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.651651 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.666527 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.689025 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.704295 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.719306 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.733567 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.742521 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.742565 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.742580 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.742596 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.742606 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:17Z","lastTransitionTime":"2026-01-26T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.758971 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:17Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303506 6178 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303603 6178 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303710 6178 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303371 6178 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 09:08:17.304122 6178 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.305297 6178 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.305440 6178 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 09:08:17.305524 6178 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 09:08:17.306178 6178 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.771516 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:17Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.845099 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.845154 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.845169 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.845192 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.845206 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:17Z","lastTransitionTime":"2026-01-26T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.948304 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.948346 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.948356 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.948372 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:17 crc kubenswrapper[4872]: I0126 09:08:17.948381 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:17Z","lastTransitionTime":"2026-01-26T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.050923 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.050967 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.050975 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.050991 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.051002 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.136595 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 23:10:32.74326197 +0000 UTC Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.154176 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.154230 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.154242 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.154259 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.154268 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.257408 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.257437 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.257447 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.257460 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.257469 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.329157 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.329198 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.329207 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.329222 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.329233 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: E0126 09:08:18.342672 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.347125 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.347175 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.347184 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.347201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.347211 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: E0126 09:08:18.360310 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.364877 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.364925 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.364938 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.364956 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.364968 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: E0126 09:08:18.376446 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.381026 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.381076 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.381089 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.381107 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.381119 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: E0126 09:08:18.400003 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.404078 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.404120 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.404130 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.404146 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.404156 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: E0126 09:08:18.416716 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: E0126 09:08:18.416895 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.418416 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.418481 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.418497 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.418517 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.418531 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.504677 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovnkube-controller/0.log" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.507960 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8"} Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.508058 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.521033 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.521077 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.521087 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.521105 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.521118 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.535085 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:17Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303506 6178 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303603 6178 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303710 6178 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303371 6178 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 09:08:17.304122 6178 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.305297 6178 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.305440 6178 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 09:08:17.305524 6178 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 09:08:17.306178 6178 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.550167 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.569572 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.591058 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.609674 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.624815 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.624870 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.624881 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.624898 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.624908 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.633413 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.648167 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.683057 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.707300 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.722932 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.728524 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.728566 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.728575 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.728591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.728603 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.743182 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.756540 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.780424 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.796489 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.811163 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:18Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.831193 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.831230 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.831247 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.831269 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.831285 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.934932 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.935002 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.935024 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.935052 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:18 crc kubenswrapper[4872]: I0126 09:08:18.935069 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:18Z","lastTransitionTime":"2026-01-26T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.038462 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.038545 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.038567 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.038596 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.038616 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:19Z","lastTransitionTime":"2026-01-26T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.136763 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 17:47:50.527308015 +0000 UTC Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.142078 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.142181 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.142208 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.142242 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.142267 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:19Z","lastTransitionTime":"2026-01-26T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.183892 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.184009 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:19 crc kubenswrapper[4872]: E0126 09:08:19.184174 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.184196 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:19 crc kubenswrapper[4872]: E0126 09:08:19.184304 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:19 crc kubenswrapper[4872]: E0126 09:08:19.184432 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.245418 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.245498 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.245510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.245531 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.245546 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:19Z","lastTransitionTime":"2026-01-26T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.348851 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.349006 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.349035 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.349065 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.349089 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:19Z","lastTransitionTime":"2026-01-26T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.451532 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.451604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.451619 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.451640 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.451657 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:19Z","lastTransitionTime":"2026-01-26T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.515726 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovnkube-controller/1.log" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.516756 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovnkube-controller/0.log" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.520833 4872 generic.go:334] "Generic (PLEG): container finished" podID="e2577331-1886-4abb-896b-8d8a112e40d8" containerID="40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8" exitCode=1 Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.520900 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8"} Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.520968 4872 scope.go:117] "RemoveContainer" containerID="3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.521952 4872 scope.go:117] "RemoveContainer" containerID="40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8" Jan 26 09:08:19 crc kubenswrapper[4872]: E0126 09:08:19.522222 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6p5bj_openshift-ovn-kubernetes(e2577331-1886-4abb-896b-8d8a112e40d8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.555379 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.555443 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.555463 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.555489 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.555509 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:19Z","lastTransitionTime":"2026-01-26T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.558503 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:17Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303506 6178 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303603 6178 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303710 6178 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303371 6178 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 09:08:17.304122 6178 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.305297 6178 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.305440 6178 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 09:08:17.305524 6178 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 09:08:17.306178 6178 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.560257 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz"] Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.561107 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.563688 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.564930 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.577626 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.603172 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.639237 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.657112 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.659492 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.659562 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.659582 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.659609 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.659627 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:19Z","lastTransitionTime":"2026-01-26T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.662570 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/87729ec8-51c9-4c4f-b49c-d1201b96939f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8h9pz\" (UID: \"87729ec8-51c9-4c4f-b49c-d1201b96939f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.662627 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/87729ec8-51c9-4c4f-b49c-d1201b96939f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8h9pz\" (UID: \"87729ec8-51c9-4c4f-b49c-d1201b96939f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.662654 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/87729ec8-51c9-4c4f-b49c-d1201b96939f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8h9pz\" (UID: \"87729ec8-51c9-4c4f-b49c-d1201b96939f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.662675 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv8m4\" (UniqueName: \"kubernetes.io/projected/87729ec8-51c9-4c4f-b49c-d1201b96939f-kube-api-access-jv8m4\") pod \"ovnkube-control-plane-749d76644c-8h9pz\" (UID: \"87729ec8-51c9-4c4f-b49c-d1201b96939f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.683364 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.704524 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.731999 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.750275 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.762540 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.762591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.762611 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.762640 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.762660 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:19Z","lastTransitionTime":"2026-01-26T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.764363 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/87729ec8-51c9-4c4f-b49c-d1201b96939f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8h9pz\" (UID: \"87729ec8-51c9-4c4f-b49c-d1201b96939f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.764471 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv8m4\" (UniqueName: \"kubernetes.io/projected/87729ec8-51c9-4c4f-b49c-d1201b96939f-kube-api-access-jv8m4\") pod \"ovnkube-control-plane-749d76644c-8h9pz\" (UID: \"87729ec8-51c9-4c4f-b49c-d1201b96939f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.764614 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/87729ec8-51c9-4c4f-b49c-d1201b96939f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8h9pz\" (UID: \"87729ec8-51c9-4c4f-b49c-d1201b96939f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.764683 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/87729ec8-51c9-4c4f-b49c-d1201b96939f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8h9pz\" (UID: \"87729ec8-51c9-4c4f-b49c-d1201b96939f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.766320 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/87729ec8-51c9-4c4f-b49c-d1201b96939f-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-8h9pz\" (UID: \"87729ec8-51c9-4c4f-b49c-d1201b96939f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.767591 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/87729ec8-51c9-4c4f-b49c-d1201b96939f-env-overrides\") pod \"ovnkube-control-plane-749d76644c-8h9pz\" (UID: \"87729ec8-51c9-4c4f-b49c-d1201b96939f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.772993 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.780740 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/87729ec8-51c9-4c4f-b49c-d1201b96939f-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-8h9pz\" (UID: \"87729ec8-51c9-4c4f-b49c-d1201b96939f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.794503 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.797974 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv8m4\" (UniqueName: \"kubernetes.io/projected/87729ec8-51c9-4c4f-b49c-d1201b96939f-kube-api-access-jv8m4\") pod \"ovnkube-control-plane-749d76644c-8h9pz\" (UID: \"87729ec8-51c9-4c4f-b49c-d1201b96939f\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.822792 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.844194 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.861159 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.865688 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.865746 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.865766 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.865794 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.865842 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:19Z","lastTransitionTime":"2026-01-26T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.877414 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.894012 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.910519 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.926776 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: W0126 09:08:19.927945 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87729ec8_51c9_4c4f_b49c_d1201b96939f.slice/crio-692f41ccec68392ac3c1ef761e6af21e73bd6ab5f1df607f3007edf0fa46584d WatchSource:0}: Error finding container 692f41ccec68392ac3c1ef761e6af21e73bd6ab5f1df607f3007edf0fa46584d: Status 404 returned error can't find the container with id 692f41ccec68392ac3c1ef761e6af21e73bd6ab5f1df607f3007edf0fa46584d Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.952976 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.968300 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.968356 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.968368 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.968434 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.968448 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:19Z","lastTransitionTime":"2026-01-26T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.971125 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:19 crc kubenswrapper[4872]: I0126 09:08:19.989297 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:19Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.004027 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.019836 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.035758 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.050695 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.064688 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.072099 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.072161 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.072181 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.072207 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.072227 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:20Z","lastTransitionTime":"2026-01-26T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.113462 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3f9a60df2ce03d7f0314118f7cb7dcf0bc10f8c410ee696aa97b125369a15b74\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:17Z\\\",\\\"message\\\":\\\"io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303506 6178 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303603 6178 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303710 6178 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.303371 6178 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI0126 09:08:17.304122 6178 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.305297 6178 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0126 09:08:17.305440 6178 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 09:08:17.305524 6178 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI0126 09:08:17.306178 6178 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.137188 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 12:15:45.305302946 +0000 UTC Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.156948 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.177708 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.177747 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.177754 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.177769 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.177779 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:20Z","lastTransitionTime":"2026-01-26T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.195401 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.210431 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.225563 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.241998 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.280499 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.281607 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.281660 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.281689 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.281737 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:20Z","lastTransitionTime":"2026-01-26T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.385572 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.385665 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.385679 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.385701 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.385720 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:20Z","lastTransitionTime":"2026-01-26T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.488663 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.488728 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.488741 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.488762 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.488776 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:20Z","lastTransitionTime":"2026-01-26T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.526189 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" event={"ID":"87729ec8-51c9-4c4f-b49c-d1201b96939f","Type":"ContainerStarted","Data":"b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f"} Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.526254 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" event={"ID":"87729ec8-51c9-4c4f-b49c-d1201b96939f","Type":"ContainerStarted","Data":"692f41ccec68392ac3c1ef761e6af21e73bd6ab5f1df607f3007edf0fa46584d"} Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.528510 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovnkube-controller/1.log" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.537438 4872 scope.go:117] "RemoveContainer" containerID="40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8" Jan 26 09:08:20 crc kubenswrapper[4872]: E0126 09:08:20.537672 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6p5bj_openshift-ovn-kubernetes(e2577331-1886-4abb-896b-8d8a112e40d8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.569458 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6p5bj_openshift-ovn-kubernetes(e2577331-1886-4abb-896b-8d8a112e40d8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.584986 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.591581 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.591648 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.591668 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.591697 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.591719 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:20Z","lastTransitionTime":"2026-01-26T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.610439 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.626009 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.640978 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.655079 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.668988 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.690580 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.694431 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.694472 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.694486 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.694502 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.694513 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:20Z","lastTransitionTime":"2026-01-26T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.705461 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.725704 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.742648 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.756559 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.775558 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.793836 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.796858 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.796922 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.796939 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.796956 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.796970 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:20Z","lastTransitionTime":"2026-01-26T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.806874 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.818607 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:20Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.899726 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.899832 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.899849 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.899873 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.899891 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:20Z","lastTransitionTime":"2026-01-26T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:20 crc kubenswrapper[4872]: I0126 09:08:20.976693 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:20 crc kubenswrapper[4872]: E0126 09:08:20.976970 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:08:36.976940332 +0000 UTC m=+50.285780153 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.003339 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.003409 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.003437 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.003471 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.003497 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:21Z","lastTransitionTime":"2026-01-26T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.071988 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-krd2b"] Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.072767 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.072917 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.077636 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.077694 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.077756 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.077790 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.077945 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.077989 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.077946 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.078074 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.078118 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.078145 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.078089 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:37.078033429 +0000 UTC m=+50.386873260 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.078080 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.078236 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.078255 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:37.078218054 +0000 UTC m=+50.387057885 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.078297 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:37.078277575 +0000 UTC m=+50.387117516 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.078340 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:37.078325736 +0000 UTC m=+50.387165667 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.100125 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.106446 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.106519 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.106544 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.106577 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.106602 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:21Z","lastTransitionTime":"2026-01-26T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.122946 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.138302 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 14:54:46.271322502 +0000 UTC Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.147615 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.163387 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.179341 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2whh7\" (UniqueName: \"kubernetes.io/projected/8f34a04d-444c-4640-9b51-51850a6317e6-kube-api-access-2whh7\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.179477 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.182257 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.183524 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.183535 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.183683 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.183897 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.184085 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.184256 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.208885 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.208965 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.208982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.209058 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.209111 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:21Z","lastTransitionTime":"2026-01-26T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.213838 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.229393 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.246360 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.265233 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.280663 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2whh7\" (UniqueName: \"kubernetes.io/projected/8f34a04d-444c-4640-9b51-51850a6317e6-kube-api-access-2whh7\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.280729 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.280850 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.280907 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs podName:8f34a04d-444c-4640-9b51-51850a6317e6 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:21.780891611 +0000 UTC m=+35.089731422 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs") pod "network-metrics-daemon-krd2b" (UID: "8f34a04d-444c-4640-9b51-51850a6317e6") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.280997 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.300320 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.301725 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2whh7\" (UniqueName: \"kubernetes.io/projected/8f34a04d-444c-4640-9b51-51850a6317e6-kube-api-access-2whh7\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.311975 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.312044 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.312061 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.312086 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.312103 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:21Z","lastTransitionTime":"2026-01-26T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.319867 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.337239 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.352350 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.371401 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.403239 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6p5bj_openshift-ovn-kubernetes(e2577331-1886-4abb-896b-8d8a112e40d8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.416065 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.416101 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.416111 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.416126 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.416135 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:21Z","lastTransitionTime":"2026-01-26T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.418915 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.518646 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.519034 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.519175 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.519321 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.519467 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:21Z","lastTransitionTime":"2026-01-26T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.542852 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" event={"ID":"87729ec8-51c9-4c4f-b49c-d1201b96939f","Type":"ContainerStarted","Data":"546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722"} Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.569310 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.593089 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.618548 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.623861 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.623912 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.623928 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.623948 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.623961 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:21Z","lastTransitionTime":"2026-01-26T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.647981 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.666232 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.700763 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.727554 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.727718 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.727846 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.727874 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.727908 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.727933 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:21Z","lastTransitionTime":"2026-01-26T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.751333 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.771506 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.787604 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.787898 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:21 crc kubenswrapper[4872]: E0126 09:08:21.788014 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs podName:8f34a04d-444c-4640-9b51-51850a6317e6 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:22.787980939 +0000 UTC m=+36.096820780 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs") pod "network-metrics-daemon-krd2b" (UID: "8f34a04d-444c-4640-9b51-51850a6317e6") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.790761 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.808908 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.830351 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.831605 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.831672 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.831684 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.831707 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.831726 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:21Z","lastTransitionTime":"2026-01-26T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.851593 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.872736 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.895274 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.928993 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6p5bj_openshift-ovn-kubernetes(e2577331-1886-4abb-896b-8d8a112e40d8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.935050 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.935093 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.935103 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.935120 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.935134 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:21Z","lastTransitionTime":"2026-01-26T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:21 crc kubenswrapper[4872]: I0126 09:08:21.949707 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:21Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.038568 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.038664 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.038683 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.038715 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.038734 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:22Z","lastTransitionTime":"2026-01-26T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.138830 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 09:44:36.486225475 +0000 UTC Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.142025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.142131 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.142159 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.142193 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.142216 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:22Z","lastTransitionTime":"2026-01-26T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.245933 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.246101 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.246136 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.246167 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.246235 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:22Z","lastTransitionTime":"2026-01-26T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.349011 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.349076 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.349094 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.349120 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.349146 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:22Z","lastTransitionTime":"2026-01-26T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.452625 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.452688 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.452702 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.452724 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.452740 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:22Z","lastTransitionTime":"2026-01-26T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.555759 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.555859 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.555878 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.555903 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.555921 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:22Z","lastTransitionTime":"2026-01-26T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.658697 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.658769 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.658789 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.658846 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.658867 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:22Z","lastTransitionTime":"2026-01-26T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.762883 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.762964 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.762987 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.763022 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.763046 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:22Z","lastTransitionTime":"2026-01-26T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.805629 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:22 crc kubenswrapper[4872]: E0126 09:08:22.805927 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:22 crc kubenswrapper[4872]: E0126 09:08:22.806052 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs podName:8f34a04d-444c-4640-9b51-51850a6317e6 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:24.806021192 +0000 UTC m=+38.114861033 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs") pod "network-metrics-daemon-krd2b" (UID: "8f34a04d-444c-4640-9b51-51850a6317e6") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.866156 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.866219 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.866231 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.866250 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.866263 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:22Z","lastTransitionTime":"2026-01-26T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.969771 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.969876 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.969897 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.969924 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:22 crc kubenswrapper[4872]: I0126 09:08:22.969943 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:22Z","lastTransitionTime":"2026-01-26T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.073662 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.073722 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.073738 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.073763 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.073779 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:23Z","lastTransitionTime":"2026-01-26T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.118770 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.120345 4872 scope.go:117] "RemoveContainer" containerID="40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8" Jan 26 09:08:23 crc kubenswrapper[4872]: E0126 09:08:23.120779 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6p5bj_openshift-ovn-kubernetes(e2577331-1886-4abb-896b-8d8a112e40d8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.139052 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 12:00:53.313665752 +0000 UTC Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.176637 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.176687 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.176700 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.176718 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.176728 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:23Z","lastTransitionTime":"2026-01-26T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.183232 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.183239 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.183244 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:23 crc kubenswrapper[4872]: E0126 09:08:23.183443 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:23 crc kubenswrapper[4872]: E0126 09:08:23.183480 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.183521 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:23 crc kubenswrapper[4872]: E0126 09:08:23.183530 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:23 crc kubenswrapper[4872]: E0126 09:08:23.183638 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.279727 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.279778 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.279790 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.279826 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.279837 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:23Z","lastTransitionTime":"2026-01-26T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.383248 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.383323 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.383347 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.383375 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.383394 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:23Z","lastTransitionTime":"2026-01-26T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.486768 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.486878 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.486893 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.486921 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.486940 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:23Z","lastTransitionTime":"2026-01-26T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.590355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.590431 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.590452 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.590481 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.590501 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:23Z","lastTransitionTime":"2026-01-26T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.694243 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.694321 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.694341 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.694371 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.694392 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:23Z","lastTransitionTime":"2026-01-26T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.797710 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.797773 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.797786 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.797820 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.797835 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:23Z","lastTransitionTime":"2026-01-26T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.900959 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.901028 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.901049 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.901077 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:23 crc kubenswrapper[4872]: I0126 09:08:23.901097 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:23Z","lastTransitionTime":"2026-01-26T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.004595 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.004649 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.004664 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.004689 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.004705 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:24Z","lastTransitionTime":"2026-01-26T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.107624 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.107678 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.107694 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.107719 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.107741 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:24Z","lastTransitionTime":"2026-01-26T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.139874 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 18:14:23.47245355 +0000 UTC Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.211082 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.211157 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.211174 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.211202 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.211220 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:24Z","lastTransitionTime":"2026-01-26T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.313658 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.313710 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.313720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.313736 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.313746 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:24Z","lastTransitionTime":"2026-01-26T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.417470 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.417536 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.417547 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.417567 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.417581 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:24Z","lastTransitionTime":"2026-01-26T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.521245 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.521374 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.521405 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.521442 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.521474 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:24Z","lastTransitionTime":"2026-01-26T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.624950 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.625023 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.625044 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.625069 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.625089 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:24Z","lastTransitionTime":"2026-01-26T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.728262 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.728376 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.728405 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.728434 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.728452 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:24Z","lastTransitionTime":"2026-01-26T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.827360 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:24 crc kubenswrapper[4872]: E0126 09:08:24.827558 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:24 crc kubenswrapper[4872]: E0126 09:08:24.827669 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs podName:8f34a04d-444c-4640-9b51-51850a6317e6 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:28.827632727 +0000 UTC m=+42.136472568 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs") pod "network-metrics-daemon-krd2b" (UID: "8f34a04d-444c-4640-9b51-51850a6317e6") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.833061 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.833131 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.833156 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.833186 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.833209 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:24Z","lastTransitionTime":"2026-01-26T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.936269 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.936349 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.936374 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.936408 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:24 crc kubenswrapper[4872]: I0126 09:08:24.936433 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:24Z","lastTransitionTime":"2026-01-26T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.039326 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.039385 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.039401 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.039424 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.039442 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:25Z","lastTransitionTime":"2026-01-26T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.140079 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 22:13:59.652540039 +0000 UTC Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.142517 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.142577 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.142594 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.142618 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.142639 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:25Z","lastTransitionTime":"2026-01-26T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.183336 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.183351 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.183369 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:25 crc kubenswrapper[4872]: E0126 09:08:25.183550 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.183597 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:25 crc kubenswrapper[4872]: E0126 09:08:25.183789 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:25 crc kubenswrapper[4872]: E0126 09:08:25.183978 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:25 crc kubenswrapper[4872]: E0126 09:08:25.184130 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.246248 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.246319 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.246337 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.246363 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.246387 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:25Z","lastTransitionTime":"2026-01-26T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.349834 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.349908 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.349925 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.349952 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.349970 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:25Z","lastTransitionTime":"2026-01-26T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.453207 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.453279 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.453295 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.453320 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.453340 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:25Z","lastTransitionTime":"2026-01-26T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.556524 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.556649 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.556672 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.556702 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.556730 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:25Z","lastTransitionTime":"2026-01-26T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.659441 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.659509 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.659537 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.659564 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.659587 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:25Z","lastTransitionTime":"2026-01-26T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.762737 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.762871 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.762897 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.762926 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.762942 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:25Z","lastTransitionTime":"2026-01-26T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.866621 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.866739 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.866758 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.866781 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.866825 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:25Z","lastTransitionTime":"2026-01-26T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.969283 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.969348 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.969362 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.969382 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:25 crc kubenswrapper[4872]: I0126 09:08:25.969393 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:25Z","lastTransitionTime":"2026-01-26T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.072403 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.072468 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.072504 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.072539 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.072565 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:26Z","lastTransitionTime":"2026-01-26T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.140325 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-14 03:17:36.133775062 +0000 UTC Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.176457 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.176542 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.176555 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.176577 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.176600 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:26Z","lastTransitionTime":"2026-01-26T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.279245 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.279311 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.279323 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.279346 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.279361 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:26Z","lastTransitionTime":"2026-01-26T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.382497 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.382570 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.382587 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.382614 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.382640 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:26Z","lastTransitionTime":"2026-01-26T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.486118 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.486191 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.486208 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.486234 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.486253 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:26Z","lastTransitionTime":"2026-01-26T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.590279 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.590383 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.590420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.590462 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.590487 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:26Z","lastTransitionTime":"2026-01-26T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.694394 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.694440 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.694449 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.694467 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.694480 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:26Z","lastTransitionTime":"2026-01-26T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.797431 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.797489 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.797506 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.797530 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.797546 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:26Z","lastTransitionTime":"2026-01-26T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.902141 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.902223 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.902247 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.902279 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:26 crc kubenswrapper[4872]: I0126 09:08:26.902304 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:26Z","lastTransitionTime":"2026-01-26T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.006059 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.006114 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.006136 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.006165 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.006186 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:27Z","lastTransitionTime":"2026-01-26T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.112016 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.112310 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.112336 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.112354 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.112378 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:27Z","lastTransitionTime":"2026-01-26T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.140547 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 13:47:39.800690534 +0000 UTC Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.184325 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.184373 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:27 crc kubenswrapper[4872]: E0126 09:08:27.184671 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.184763 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.184876 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:27 crc kubenswrapper[4872]: E0126 09:08:27.184953 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:27 crc kubenswrapper[4872]: E0126 09:08:27.185116 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:27 crc kubenswrapper[4872]: E0126 09:08:27.185514 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.208120 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.216159 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.216223 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.216249 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.216278 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.216301 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:27Z","lastTransitionTime":"2026-01-26T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.242887 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.256605 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.269280 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.280963 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.292141 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.304849 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.316527 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.318544 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.318583 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.318591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.318606 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.318616 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:27Z","lastTransitionTime":"2026-01-26T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.328583 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.338588 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.348705 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.372876 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6p5bj_openshift-ovn-kubernetes(e2577331-1886-4abb-896b-8d8a112e40d8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.383499 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.400046 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.417435 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.421415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.421454 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.421465 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.421479 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.421491 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:27Z","lastTransitionTime":"2026-01-26T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.431147 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.444989 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:27Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.524617 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.524676 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.524694 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.524720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.524737 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:27Z","lastTransitionTime":"2026-01-26T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.629334 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.629420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.629438 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.629467 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.629495 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:27Z","lastTransitionTime":"2026-01-26T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.733226 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.733269 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.733279 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.733296 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.733307 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:27Z","lastTransitionTime":"2026-01-26T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.836333 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.837364 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.837475 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.837835 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.838350 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:27Z","lastTransitionTime":"2026-01-26T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.941900 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.941958 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.941970 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.941992 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:27 crc kubenswrapper[4872]: I0126 09:08:27.942010 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:27Z","lastTransitionTime":"2026-01-26T09:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.045871 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.045974 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.045993 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.046054 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.046075 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.141518 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 19:11:30.317012775 +0000 UTC Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.149138 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.149243 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.149263 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.149293 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.149311 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.252686 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.252750 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.252762 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.252783 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.252818 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.356423 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.356497 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.356509 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.356531 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.356549 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.459610 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.459672 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.459687 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.459715 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.459733 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.562849 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.562929 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.562948 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.562972 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.562989 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.666939 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.666989 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.666998 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.667018 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.667034 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.698350 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.698413 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.698423 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.698443 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.698459 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: E0126 09:08:28.716566 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:28Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.722169 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.722278 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.722297 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.722359 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.722379 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: E0126 09:08:28.742603 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:28Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.749053 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.749088 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.749097 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.749114 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.749127 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: E0126 09:08:28.763424 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:28Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.768090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.768124 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.768134 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.768150 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.768161 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: E0126 09:08:28.781972 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:28Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.786528 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.786570 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.786580 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.786599 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.786631 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: E0126 09:08:28.809250 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:28Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:28 crc kubenswrapper[4872]: E0126 09:08:28.809465 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.812028 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.812107 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.812121 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.812143 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.812159 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.879384 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:28 crc kubenswrapper[4872]: E0126 09:08:28.879594 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:28 crc kubenswrapper[4872]: E0126 09:08:28.879675 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs podName:8f34a04d-444c-4640-9b51-51850a6317e6 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:36.879655131 +0000 UTC m=+50.188494952 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs") pod "network-metrics-daemon-krd2b" (UID: "8f34a04d-444c-4640-9b51-51850a6317e6") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.915649 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.915734 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.915751 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.915779 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:28 crc kubenswrapper[4872]: I0126 09:08:28.915832 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:28Z","lastTransitionTime":"2026-01-26T09:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.019547 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.019627 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.019646 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.019675 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.019696 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:29Z","lastTransitionTime":"2026-01-26T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.123418 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.123484 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.123496 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.123521 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.123534 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:29Z","lastTransitionTime":"2026-01-26T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.142106 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 19:25:36.082300181 +0000 UTC Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.183709 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.183821 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.183899 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:29 crc kubenswrapper[4872]: E0126 09:08:29.183934 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.183732 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:29 crc kubenswrapper[4872]: E0126 09:08:29.184128 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:29 crc kubenswrapper[4872]: E0126 09:08:29.184225 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:29 crc kubenswrapper[4872]: E0126 09:08:29.184399 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.226943 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.227003 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.227019 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.227046 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.227065 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:29Z","lastTransitionTime":"2026-01-26T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.331710 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.331785 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.331835 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.331872 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.331901 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:29Z","lastTransitionTime":"2026-01-26T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.434771 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.434842 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.434855 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.434874 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.434886 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:29Z","lastTransitionTime":"2026-01-26T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.538674 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.538768 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.538794 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.538916 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.538948 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:29Z","lastTransitionTime":"2026-01-26T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.642183 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.642255 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.642274 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.642304 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.642325 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:29Z","lastTransitionTime":"2026-01-26T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.746355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.746417 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.746430 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.746454 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.746474 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:29Z","lastTransitionTime":"2026-01-26T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.849376 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.849429 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.849442 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.849463 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.849493 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:29Z","lastTransitionTime":"2026-01-26T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.952372 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.952429 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.952444 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.952466 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:29 crc kubenswrapper[4872]: I0126 09:08:29.952482 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:29Z","lastTransitionTime":"2026-01-26T09:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.055380 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.055487 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.055500 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.055528 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.055541 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:30Z","lastTransitionTime":"2026-01-26T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.142464 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 01:49:36.502092662 +0000 UTC Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.159206 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.159324 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.159362 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.159396 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.159421 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:30Z","lastTransitionTime":"2026-01-26T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.263123 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.263221 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.263246 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.263278 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.263298 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:30Z","lastTransitionTime":"2026-01-26T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.366440 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.366523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.366548 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.366581 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.366602 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:30Z","lastTransitionTime":"2026-01-26T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.469827 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.469935 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.469963 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.469995 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.470017 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:30Z","lastTransitionTime":"2026-01-26T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.573607 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.573673 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.573696 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.573727 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.573751 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:30Z","lastTransitionTime":"2026-01-26T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.676894 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.676962 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.677006 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.677032 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.677050 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:30Z","lastTransitionTime":"2026-01-26T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.780833 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.780912 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.780932 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.780959 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.780979 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:30Z","lastTransitionTime":"2026-01-26T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.891449 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.891504 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.891522 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.891550 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.891570 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:30Z","lastTransitionTime":"2026-01-26T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.995655 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.995718 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.995738 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.995768 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:30 crc kubenswrapper[4872]: I0126 09:08:30.995793 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:30Z","lastTransitionTime":"2026-01-26T09:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.098930 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.098997 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.099043 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.099072 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.099096 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:31Z","lastTransitionTime":"2026-01-26T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.142855 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 13:48:48.569236332 +0000 UTC Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.183498 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.183549 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:31 crc kubenswrapper[4872]: E0126 09:08:31.183949 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.184069 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:31 crc kubenswrapper[4872]: E0126 09:08:31.184210 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.184314 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:31 crc kubenswrapper[4872]: E0126 09:08:31.184568 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:31 crc kubenswrapper[4872]: E0126 09:08:31.184747 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.202520 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.202598 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.202618 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.202649 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.202671 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:31Z","lastTransitionTime":"2026-01-26T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.306392 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.306451 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.306461 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.306485 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.306498 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:31Z","lastTransitionTime":"2026-01-26T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.409959 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.410018 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.410027 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.410050 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.410061 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:31Z","lastTransitionTime":"2026-01-26T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.512779 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.512931 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.512951 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.512974 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.512991 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:31Z","lastTransitionTime":"2026-01-26T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.615490 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.615555 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.615572 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.615590 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.615602 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:31Z","lastTransitionTime":"2026-01-26T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.719020 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.719103 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.719126 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.719156 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.719179 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:31Z","lastTransitionTime":"2026-01-26T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.823013 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.823066 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.823078 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.823097 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.823109 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:31Z","lastTransitionTime":"2026-01-26T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.926380 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.926452 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.926471 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.926495 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:31 crc kubenswrapper[4872]: I0126 09:08:31.926513 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:31Z","lastTransitionTime":"2026-01-26T09:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.029893 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.029950 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.029963 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.029982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.029995 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:32Z","lastTransitionTime":"2026-01-26T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.133007 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.133090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.133101 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.133119 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.133130 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:32Z","lastTransitionTime":"2026-01-26T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.143305 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 09:30:14.511083064 +0000 UTC Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.236235 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.236300 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.236314 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.236332 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.236343 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:32Z","lastTransitionTime":"2026-01-26T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.339983 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.340037 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.340052 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.340076 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.340092 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:32Z","lastTransitionTime":"2026-01-26T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.443240 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.443283 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.443294 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.443310 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.443322 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:32Z","lastTransitionTime":"2026-01-26T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.546571 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.547000 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.547088 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.547157 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.547247 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:32Z","lastTransitionTime":"2026-01-26T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.650234 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.650302 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.650316 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.650339 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.650354 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:32Z","lastTransitionTime":"2026-01-26T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.753976 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.754070 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.754098 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.754131 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.754156 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:32Z","lastTransitionTime":"2026-01-26T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.858370 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.858449 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.858468 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.858497 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.858516 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:32Z","lastTransitionTime":"2026-01-26T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.961666 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.961779 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.961857 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.961882 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:32 crc kubenswrapper[4872]: I0126 09:08:32.961931 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:32Z","lastTransitionTime":"2026-01-26T09:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.065704 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.065785 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.065852 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.065881 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.065904 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:33Z","lastTransitionTime":"2026-01-26T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.144182 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 01:01:28.416546736 +0000 UTC Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.168676 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.168829 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.168851 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.168875 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.168924 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:33Z","lastTransitionTime":"2026-01-26T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.183537 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.183658 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:33 crc kubenswrapper[4872]: E0126 09:08:33.183740 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.183572 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.183980 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:33 crc kubenswrapper[4872]: E0126 09:08:33.184118 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:33 crc kubenswrapper[4872]: E0126 09:08:33.185051 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:33 crc kubenswrapper[4872]: E0126 09:08:33.185137 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.272756 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.272829 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.272840 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.272859 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.272871 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:33Z","lastTransitionTime":"2026-01-26T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.376252 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.376314 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.376324 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.376353 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.376365 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:33Z","lastTransitionTime":"2026-01-26T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.479477 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.479532 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.479541 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.479753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.479765 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:33Z","lastTransitionTime":"2026-01-26T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.583723 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.583825 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.583837 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.583855 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.583868 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:33Z","lastTransitionTime":"2026-01-26T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.686781 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.686876 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.686956 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.686982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.687025 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:33Z","lastTransitionTime":"2026-01-26T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.790357 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.790443 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.790469 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.790498 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.790516 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:33Z","lastTransitionTime":"2026-01-26T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.893705 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.893774 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.893831 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.893864 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.893886 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:33Z","lastTransitionTime":"2026-01-26T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.998235 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.998302 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.998318 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.998341 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:33 crc kubenswrapper[4872]: I0126 09:08:33.998356 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:33Z","lastTransitionTime":"2026-01-26T09:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.102117 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.102187 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.102201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.102222 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.102236 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:34Z","lastTransitionTime":"2026-01-26T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.144389 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 18:43:39.188921741 +0000 UTC Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.206447 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.206525 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.206543 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.206572 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.206591 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:34Z","lastTransitionTime":"2026-01-26T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.310370 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.310484 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.310510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.310541 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.310563 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:34Z","lastTransitionTime":"2026-01-26T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.413844 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.413907 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.413922 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.413947 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.413964 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:34Z","lastTransitionTime":"2026-01-26T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.517273 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.517334 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.517348 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.517371 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.517386 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:34Z","lastTransitionTime":"2026-01-26T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.620001 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.620074 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.620092 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.620118 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.620136 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:34Z","lastTransitionTime":"2026-01-26T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.722998 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.723059 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.723077 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.723100 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.723118 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:34Z","lastTransitionTime":"2026-01-26T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.826216 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.826309 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.826334 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.826373 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.826400 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:34Z","lastTransitionTime":"2026-01-26T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.929864 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.929949 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.929973 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.930017 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:34 crc kubenswrapper[4872]: I0126 09:08:34.930038 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:34Z","lastTransitionTime":"2026-01-26T09:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.033955 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.034028 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.034047 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.034076 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.034098 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:35Z","lastTransitionTime":"2026-01-26T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.137184 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.137254 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.137272 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.137299 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.137320 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:35Z","lastTransitionTime":"2026-01-26T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.145394 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-24 21:49:49.679462352 +0000 UTC Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.183068 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.183137 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:35 crc kubenswrapper[4872]: E0126 09:08:35.183256 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.183310 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.183408 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:35 crc kubenswrapper[4872]: E0126 09:08:35.183624 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:35 crc kubenswrapper[4872]: E0126 09:08:35.183766 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:35 crc kubenswrapper[4872]: E0126 09:08:35.184021 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.241166 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.241223 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.241232 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.241250 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.241261 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:35Z","lastTransitionTime":"2026-01-26T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.343876 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.343949 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.343970 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.344000 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.344020 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:35Z","lastTransitionTime":"2026-01-26T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.447064 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.447118 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.447133 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.447156 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.447173 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:35Z","lastTransitionTime":"2026-01-26T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.550551 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.550604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.550617 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.550637 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.550650 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:35Z","lastTransitionTime":"2026-01-26T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.654392 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.654450 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.654466 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.654489 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.654507 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:35Z","lastTransitionTime":"2026-01-26T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.758029 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.758117 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.758141 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.758201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.758225 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:35Z","lastTransitionTime":"2026-01-26T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.861728 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.861791 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.861844 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.861868 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.861889 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:35Z","lastTransitionTime":"2026-01-26T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.965543 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.965628 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.965665 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.965698 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:35 crc kubenswrapper[4872]: I0126 09:08:35.965720 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:35Z","lastTransitionTime":"2026-01-26T09:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.068739 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.068846 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.068865 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.068891 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.068913 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:36Z","lastTransitionTime":"2026-01-26T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.146039 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 01:20:16.069528009 +0000 UTC Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.172096 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.172170 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.172193 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.172224 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.172248 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:36Z","lastTransitionTime":"2026-01-26T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.275515 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.275997 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.276184 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.276320 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.276426 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:36Z","lastTransitionTime":"2026-01-26T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.379697 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.379756 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.379767 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.379791 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.379827 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:36Z","lastTransitionTime":"2026-01-26T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.483657 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.483712 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.483723 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.483744 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.483756 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:36Z","lastTransitionTime":"2026-01-26T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.587482 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.587555 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.587568 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.587590 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.587606 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:36Z","lastTransitionTime":"2026-01-26T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.692049 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.692147 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.692170 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.692200 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.692219 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:36Z","lastTransitionTime":"2026-01-26T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.795730 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.795780 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.795790 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.795829 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.795841 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:36Z","lastTransitionTime":"2026-01-26T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.898413 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.898477 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.898489 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.898510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.898522 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:36Z","lastTransitionTime":"2026-01-26T09:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.977612 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:08:36 crc kubenswrapper[4872]: E0126 09:08:36.977925 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:08.977879567 +0000 UTC m=+82.286719368 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:08:36 crc kubenswrapper[4872]: I0126 09:08:36.978018 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:36 crc kubenswrapper[4872]: E0126 09:08:36.978280 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:36 crc kubenswrapper[4872]: E0126 09:08:36.978408 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs podName:8f34a04d-444c-4640-9b51-51850a6317e6 nodeName:}" failed. No retries permitted until 2026-01-26 09:08:52.978373911 +0000 UTC m=+66.287213872 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs") pod "network-metrics-daemon-krd2b" (UID: "8f34a04d-444c-4640-9b51-51850a6317e6") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.002096 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.002191 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.002209 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.002233 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.002249 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:37Z","lastTransitionTime":"2026-01-26T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.078947 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.078998 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.079037 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.079057 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.079177 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.079224 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:09:09.0792105 +0000 UTC m=+82.388050301 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.079250 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.079352 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:09:09.079330123 +0000 UTC m=+82.388169924 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.079408 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.079422 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.079433 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.079454 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 09:09:09.079447796 +0000 UTC m=+82.388287597 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.079531 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.079543 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.079551 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.079571 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 09:09:09.07956601 +0000 UTC m=+82.388405811 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.104281 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.104329 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.104340 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.104357 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.104370 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:37Z","lastTransitionTime":"2026-01-26T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.147207 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 22:25:08.070864136 +0000 UTC Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.184040 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.184193 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.184368 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.184583 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.184635 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.184708 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.184816 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:37 crc kubenswrapper[4872]: E0126 09:08:37.185147 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.185538 4872 scope.go:117] "RemoveContainer" containerID="40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.200863 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.206941 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.206986 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.206997 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.207015 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.207028 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:37Z","lastTransitionTime":"2026-01-26T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.225915 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.240677 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.260486 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.281218 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.304415 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.309963 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.310001 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.310019 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.310040 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.310056 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:37Z","lastTransitionTime":"2026-01-26T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.320877 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.333492 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.347856 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.364104 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.384513 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.407621 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.412765 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.412826 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.412837 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.412858 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.412870 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:37Z","lastTransitionTime":"2026-01-26T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.426398 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.441003 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.453374 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.475898 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-6p5bj_openshift-ovn-kubernetes(e2577331-1886-4abb-896b-8d8a112e40d8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.490468 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.514993 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.515045 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.515053 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.515069 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.515078 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:37Z","lastTransitionTime":"2026-01-26T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.616646 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.616704 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.616719 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.616740 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.616749 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovnkube-controller/1.log" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.616756 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:37Z","lastTransitionTime":"2026-01-26T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.619981 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621"} Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.620556 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.643019 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.685227 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.685242 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.704768 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.713504 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.719130 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.719167 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.719176 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.719191 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.719204 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:37Z","lastTransitionTime":"2026-01-26T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.723087 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.742959 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.778244 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.799029 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.822006 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.822049 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.822060 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.822078 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.822091 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:37Z","lastTransitionTime":"2026-01-26T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.826627 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.842133 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.855901 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.869319 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.886609 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.903869 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.919421 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.925584 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.925626 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.925636 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.925654 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.925668 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:37Z","lastTransitionTime":"2026-01-26T09:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.935201 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.950641 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.966338 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.980987 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:37 crc kubenswrapper[4872]: I0126 09:08:37.997320 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:37Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.019011 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.028328 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.028385 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.028397 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.028417 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.028430 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:38Z","lastTransitionTime":"2026-01-26T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.036014 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.050549 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.065660 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.079010 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.130636 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.130684 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.130697 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.130715 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.130730 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:38Z","lastTransitionTime":"2026-01-26T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.138993 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.147517 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 22:15:46.829892299 +0000 UTC Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.151593 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.175724 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.186739 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.198166 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.215098 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13bb0fc1-b822-4651-ab5f-b095c86e936b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1754e95c65f24f7086f23e38595f8d3b6acb9baaaed809afee607eb5a13d06bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178c8017e99cb8783654f3249f3499f6e61c7cabfbedf1bd078e3f345b3e7cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3c7b1aa5169eb070e390f0b0e5f06149b73145a6a041e0c3e6ba46857f77814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.233652 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.233695 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.233705 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.233722 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.233736 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:38Z","lastTransitionTime":"2026-01-26T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.238226 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.251092 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.269972 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.287646 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.305204 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:38Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.336965 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.337084 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.337097 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.337120 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.337134 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:38Z","lastTransitionTime":"2026-01-26T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.439452 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.439510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.439523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.439544 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.439557 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:38Z","lastTransitionTime":"2026-01-26T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.543064 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.543112 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.543123 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.543143 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.543156 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:38Z","lastTransitionTime":"2026-01-26T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.646569 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.646895 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.647040 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.647154 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.647247 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:38Z","lastTransitionTime":"2026-01-26T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.750204 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.750271 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.750289 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.750317 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.750337 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:38Z","lastTransitionTime":"2026-01-26T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.854483 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.854543 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.854558 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.854581 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.854601 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:38Z","lastTransitionTime":"2026-01-26T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.957576 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.957632 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.957646 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.957670 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:38 crc kubenswrapper[4872]: I0126 09:08:38.957685 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:38Z","lastTransitionTime":"2026-01-26T09:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.016305 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.016405 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.016435 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.016470 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.016501 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: E0126 09:08:39.035354 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:39Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.039886 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.039931 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.039949 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.039975 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.039993 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: E0126 09:08:39.064419 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:39Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.069962 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.070026 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.070045 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.070071 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.070092 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: E0126 09:08:39.086217 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:39Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.091010 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.091059 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.091070 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.091091 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.091105 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: E0126 09:08:39.107103 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:39Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.114442 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.114586 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.114606 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.114640 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.114662 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: E0126 09:08:39.131516 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:39Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:39 crc kubenswrapper[4872]: E0126 09:08:39.131701 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.133938 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.133982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.133995 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.134015 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.134031 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.148390 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 07:06:11.882245724 +0000 UTC Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.184020 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.184020 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.184045 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:39 crc kubenswrapper[4872]: E0126 09:08:39.184183 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:39 crc kubenswrapper[4872]: E0126 09:08:39.184231 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.184263 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:39 crc kubenswrapper[4872]: E0126 09:08:39.184283 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:39 crc kubenswrapper[4872]: E0126 09:08:39.184405 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.236980 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.237372 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.237548 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.237899 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.238070 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.341155 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.341230 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.341247 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.341274 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.341293 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.445313 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.445406 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.445428 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.445460 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.445489 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.548537 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.548969 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.549179 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.549409 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.549585 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.653128 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.653209 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.653229 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.653261 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.653281 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.756920 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.756964 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.757010 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.757035 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.757083 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.860201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.860243 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.860255 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.860273 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.860287 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.963881 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.963946 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.964004 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.964027 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:39 crc kubenswrapper[4872]: I0126 09:08:39.964042 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:39Z","lastTransitionTime":"2026-01-26T09:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.067016 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.067067 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.067079 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.067097 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.067111 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:40Z","lastTransitionTime":"2026-01-26T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.149153 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 20:38:31.068323309 +0000 UTC Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.170501 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.170570 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.170588 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.170614 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.170631 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:40Z","lastTransitionTime":"2026-01-26T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.274890 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.274939 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.274990 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.275016 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.275031 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:40Z","lastTransitionTime":"2026-01-26T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.378157 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.378239 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.378260 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.378287 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.378304 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:40Z","lastTransitionTime":"2026-01-26T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.482064 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.482116 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.482141 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.482167 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.482185 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:40Z","lastTransitionTime":"2026-01-26T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.585283 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.585349 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.585371 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.585397 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.585416 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:40Z","lastTransitionTime":"2026-01-26T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.688982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.689082 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.689103 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.689137 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.689161 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:40Z","lastTransitionTime":"2026-01-26T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.792450 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.792520 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.792532 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.792551 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.792564 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:40Z","lastTransitionTime":"2026-01-26T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.896448 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.896830 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.896953 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.897052 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:40 crc kubenswrapper[4872]: I0126 09:08:40.897157 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:40Z","lastTransitionTime":"2026-01-26T09:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.000756 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.000839 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.000859 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.000885 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.000903 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:41Z","lastTransitionTime":"2026-01-26T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.104085 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.104144 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.104160 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.104185 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.104202 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:41Z","lastTransitionTime":"2026-01-26T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.149586 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 06:50:01.856550908 +0000 UTC Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.183654 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:41 crc kubenswrapper[4872]: E0126 09:08:41.183900 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.184172 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.184250 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.184183 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:41 crc kubenswrapper[4872]: E0126 09:08:41.184423 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:41 crc kubenswrapper[4872]: E0126 09:08:41.184500 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:41 crc kubenswrapper[4872]: E0126 09:08:41.184607 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.207359 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.207424 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.207442 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.207465 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.207481 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:41Z","lastTransitionTime":"2026-01-26T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.311479 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.311555 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.311580 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.311618 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.311640 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:41Z","lastTransitionTime":"2026-01-26T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.415335 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.415399 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.415425 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.415452 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.415471 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:41Z","lastTransitionTime":"2026-01-26T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.518874 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.518921 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.518937 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.518958 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.518978 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:41Z","lastTransitionTime":"2026-01-26T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.622480 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.622544 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.622555 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.622575 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.622589 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:41Z","lastTransitionTime":"2026-01-26T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.726169 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.726228 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.726242 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.726262 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.726277 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:41Z","lastTransitionTime":"2026-01-26T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.829458 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.829653 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.829925 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.829954 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.829975 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:41Z","lastTransitionTime":"2026-01-26T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.934163 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.934229 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.934241 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.934262 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:41 crc kubenswrapper[4872]: I0126 09:08:41.934274 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:41Z","lastTransitionTime":"2026-01-26T09:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.041130 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.041189 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.041201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.041218 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.041229 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:42Z","lastTransitionTime":"2026-01-26T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.144230 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.144281 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.144291 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.144309 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.144321 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:42Z","lastTransitionTime":"2026-01-26T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.149795 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 21:07:25.070926182 +0000 UTC Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.247566 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.247637 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.247649 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.247670 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.247683 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:42Z","lastTransitionTime":"2026-01-26T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.350604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.350650 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.350661 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.350679 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.350692 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:42Z","lastTransitionTime":"2026-01-26T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.453293 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.453356 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.453367 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.453388 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.453401 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:42Z","lastTransitionTime":"2026-01-26T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.556380 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.556432 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.556448 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.556470 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.556482 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:42Z","lastTransitionTime":"2026-01-26T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.658992 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.659053 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.659071 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.659123 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.659143 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:42Z","lastTransitionTime":"2026-01-26T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.762498 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.762598 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.762620 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.762650 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.762669 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:42Z","lastTransitionTime":"2026-01-26T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.865893 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.865942 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.865950 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.865971 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.865981 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:42Z","lastTransitionTime":"2026-01-26T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.968865 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.968946 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.968966 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.968992 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:42 crc kubenswrapper[4872]: I0126 09:08:42.969010 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:42Z","lastTransitionTime":"2026-01-26T09:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.072293 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.072361 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.072382 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.072406 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.072426 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:43Z","lastTransitionTime":"2026-01-26T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.151047 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 22:03:53.472325744 +0000 UTC Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.176176 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.176236 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.176253 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.176293 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.176315 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:43Z","lastTransitionTime":"2026-01-26T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.183515 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.183564 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.183687 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:43 crc kubenswrapper[4872]: E0126 09:08:43.183683 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:43 crc kubenswrapper[4872]: E0126 09:08:43.183924 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:43 crc kubenswrapper[4872]: E0126 09:08:43.184043 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.184116 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:43 crc kubenswrapper[4872]: E0126 09:08:43.184186 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.279415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.279486 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.279508 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.279531 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.279548 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:43Z","lastTransitionTime":"2026-01-26T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.382657 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.382737 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.382760 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.382788 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.382849 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:43Z","lastTransitionTime":"2026-01-26T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.485260 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.485303 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.485321 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.485342 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.485356 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:43Z","lastTransitionTime":"2026-01-26T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.587817 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.587872 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.587884 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.587904 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.587916 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:43Z","lastTransitionTime":"2026-01-26T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.691610 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.691669 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.691685 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.691709 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.691722 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:43Z","lastTransitionTime":"2026-01-26T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.794782 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.795047 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.795060 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.795081 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.795097 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:43Z","lastTransitionTime":"2026-01-26T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.898463 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.898515 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.898524 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.898542 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:43 crc kubenswrapper[4872]: I0126 09:08:43.898553 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:43Z","lastTransitionTime":"2026-01-26T09:08:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.001566 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.001630 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.001644 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.001665 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.001679 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:44Z","lastTransitionTime":"2026-01-26T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.105482 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.105537 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.105553 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.105577 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.105595 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:44Z","lastTransitionTime":"2026-01-26T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.151186 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 07:27:53.121437905 +0000 UTC Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.209318 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.209400 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.209422 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.209452 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.209477 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:44Z","lastTransitionTime":"2026-01-26T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.317019 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.317141 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.317239 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.317260 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.317275 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:44Z","lastTransitionTime":"2026-01-26T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.421118 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.421202 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.421219 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.421247 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.421268 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:44Z","lastTransitionTime":"2026-01-26T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.525132 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.525177 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.525187 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.525204 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.525217 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:44Z","lastTransitionTime":"2026-01-26T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.628878 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.628940 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.628963 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.628992 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.629014 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:44Z","lastTransitionTime":"2026-01-26T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.731969 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.732033 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.732053 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.732077 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.732098 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:44Z","lastTransitionTime":"2026-01-26T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.835648 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.835716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.835735 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.835764 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.835783 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:44Z","lastTransitionTime":"2026-01-26T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.939352 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.939426 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.939449 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.939484 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:44 crc kubenswrapper[4872]: I0126 09:08:44.939509 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:44Z","lastTransitionTime":"2026-01-26T09:08:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.042596 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.042668 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.042689 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.042717 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.042739 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:45Z","lastTransitionTime":"2026-01-26T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.146251 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.146310 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.146327 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.146352 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.146369 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:45Z","lastTransitionTime":"2026-01-26T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.151720 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 14:19:53.014031381 +0000 UTC Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.183345 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.183463 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.183478 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:45 crc kubenswrapper[4872]: E0126 09:08:45.183633 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.183676 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:45 crc kubenswrapper[4872]: E0126 09:08:45.183787 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:45 crc kubenswrapper[4872]: E0126 09:08:45.184003 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:45 crc kubenswrapper[4872]: E0126 09:08:45.184192 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.249314 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.249380 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.249401 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.249432 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.249486 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:45Z","lastTransitionTime":"2026-01-26T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.352399 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.352487 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.352510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.352539 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.352560 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:45Z","lastTransitionTime":"2026-01-26T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.455992 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.456048 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.456067 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.456090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.456107 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:45Z","lastTransitionTime":"2026-01-26T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.559670 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.559745 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.559763 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.559789 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.559844 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:45Z","lastTransitionTime":"2026-01-26T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.663302 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.663373 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.663389 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.663413 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.663431 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:45Z","lastTransitionTime":"2026-01-26T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.766681 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.766760 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.766778 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.766832 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.766853 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:45Z","lastTransitionTime":"2026-01-26T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.870262 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.870330 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.870346 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.870372 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.870391 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:45Z","lastTransitionTime":"2026-01-26T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.973015 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.973081 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.973097 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.973119 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:45 crc kubenswrapper[4872]: I0126 09:08:45.973140 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:45Z","lastTransitionTime":"2026-01-26T09:08:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.077472 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.077547 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.077574 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.077606 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.077626 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:46Z","lastTransitionTime":"2026-01-26T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.152543 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 17:10:53.051896496 +0000 UTC Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.181152 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.181205 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.181223 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.181274 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.181293 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:46Z","lastTransitionTime":"2026-01-26T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.284385 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.284488 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.284508 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.284536 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.284565 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:46Z","lastTransitionTime":"2026-01-26T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.387907 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.388029 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.388060 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.388090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.388119 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:46Z","lastTransitionTime":"2026-01-26T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.491182 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.491247 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.491255 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.491272 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.491283 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:46Z","lastTransitionTime":"2026-01-26T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.595028 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.595084 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.595095 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.595121 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.595135 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:46Z","lastTransitionTime":"2026-01-26T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.698564 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.698635 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.698653 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.698679 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.698700 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:46Z","lastTransitionTime":"2026-01-26T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.801970 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.802043 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.802061 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.802084 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.802099 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:46Z","lastTransitionTime":"2026-01-26T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.905898 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.906007 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.906023 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.906044 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:46 crc kubenswrapper[4872]: I0126 09:08:46.906060 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:46Z","lastTransitionTime":"2026-01-26T09:08:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.008977 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.009040 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.009057 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.009080 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.009100 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:47Z","lastTransitionTime":"2026-01-26T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.111863 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.111922 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.111934 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.111954 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.111966 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:47Z","lastTransitionTime":"2026-01-26T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.153623 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-11 09:07:32.103366203 +0000 UTC Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.183557 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.183603 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.183597 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.183677 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:47 crc kubenswrapper[4872]: E0126 09:08:47.183889 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:47 crc kubenswrapper[4872]: E0126 09:08:47.183991 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:47 crc kubenswrapper[4872]: E0126 09:08:47.184068 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:47 crc kubenswrapper[4872]: E0126 09:08:47.184206 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.204056 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.215093 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.215151 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.215162 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.215182 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.215197 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:47Z","lastTransitionTime":"2026-01-26T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.229255 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.244690 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.258049 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.274993 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.290897 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.305449 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.318259 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.318317 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.318327 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.318348 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.318362 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:47Z","lastTransitionTime":"2026-01-26T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.323962 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.337915 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.357047 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.369109 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.381014 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13bb0fc1-b822-4651-ab5f-b095c86e936b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1754e95c65f24f7086f23e38595f8d3b6acb9baaaed809afee607eb5a13d06bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178c8017e99cb8783654f3249f3499f6e61c7cabfbedf1bd078e3f345b3e7cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3c7b1aa5169eb070e390f0b0e5f06149b73145a6a041e0c3e6ba46857f77814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.403503 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.419048 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.420943 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.420978 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.420987 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.421003 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.421014 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:47Z","lastTransitionTime":"2026-01-26T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.437182 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.452907 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.466790 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.486696 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:47Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.523266 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.523306 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.523319 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.523337 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.523349 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:47Z","lastTransitionTime":"2026-01-26T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.627036 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.627111 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.627320 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.627351 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.627373 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:47Z","lastTransitionTime":"2026-01-26T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.730106 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.730541 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.730715 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.730945 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.731001 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:47Z","lastTransitionTime":"2026-01-26T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.834099 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.834243 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.834286 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.834338 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.834363 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:47Z","lastTransitionTime":"2026-01-26T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.937006 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.937087 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.937108 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.937135 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:47 crc kubenswrapper[4872]: I0126 09:08:47.937151 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:47Z","lastTransitionTime":"2026-01-26T09:08:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.039342 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.039401 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.039417 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.039441 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.039457 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:48Z","lastTransitionTime":"2026-01-26T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.142411 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.142459 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.142471 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.142490 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.142502 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:48Z","lastTransitionTime":"2026-01-26T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.153991 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 07:21:51.739609643 +0000 UTC Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.245545 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.245657 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.245680 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.245716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.245738 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:48Z","lastTransitionTime":"2026-01-26T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.348880 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.348927 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.348937 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.348954 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.348964 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:48Z","lastTransitionTime":"2026-01-26T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.452922 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.453414 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.453437 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.453466 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.453489 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:48Z","lastTransitionTime":"2026-01-26T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.556114 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.556233 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.556266 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.556302 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.556327 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:48Z","lastTransitionTime":"2026-01-26T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.660222 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.660297 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.660309 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.660334 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.660348 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:48Z","lastTransitionTime":"2026-01-26T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.763971 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.764019 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.764028 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.764049 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.764066 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:48Z","lastTransitionTime":"2026-01-26T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.867903 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.868266 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.868418 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.868604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.868768 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:48Z","lastTransitionTime":"2026-01-26T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.985690 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.985858 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.985893 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.985930 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:48 crc kubenswrapper[4872]: I0126 09:08:48.985965 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:48Z","lastTransitionTime":"2026-01-26T09:08:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.089927 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.090001 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.090025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.090062 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.090088 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.154786 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 09:36:31.914633322 +0000 UTC Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.163751 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.163837 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.163861 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.163889 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.163911 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.183867 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:49 crc kubenswrapper[4872]: E0126 09:08:49.184102 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.184517 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.184760 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:49 crc kubenswrapper[4872]: E0126 09:08:49.184788 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:49 crc kubenswrapper[4872]: E0126 09:08:49.185361 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.185571 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:49 crc kubenswrapper[4872]: E0126 09:08:49.185720 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:49 crc kubenswrapper[4872]: E0126 09:08:49.187195 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:49Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.196213 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.196464 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.196622 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.196767 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.196987 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: E0126 09:08:49.218359 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:49Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.223746 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.223787 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.223824 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.223848 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.223865 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: E0126 09:08:49.240355 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:49Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.244969 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.245046 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.245065 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.245099 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.245123 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: E0126 09:08:49.263560 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:49Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.268359 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.268415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.268436 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.268461 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.268479 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: E0126 09:08:49.286443 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:49Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:49 crc kubenswrapper[4872]: E0126 09:08:49.287078 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.289770 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.289879 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.289902 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.289931 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.289952 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.392939 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.393472 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.393653 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.393852 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.394053 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.497598 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.497662 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.497689 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.497720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.497743 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.601343 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.601416 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.601432 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.601463 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.601478 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.705393 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.705470 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.705490 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.705523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.705546 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.808835 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.808891 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.808906 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.808933 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.808949 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.911083 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.911133 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.911152 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.911173 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:49 crc kubenswrapper[4872]: I0126 09:08:49.911188 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:49Z","lastTransitionTime":"2026-01-26T09:08:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.014967 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.015066 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.015089 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.015126 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.015149 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:50Z","lastTransitionTime":"2026-01-26T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.118122 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.118541 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.118711 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.118880 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.119029 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:50Z","lastTransitionTime":"2026-01-26T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.156034 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-04 08:52:48.594801195 +0000 UTC Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.222702 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.222773 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.222793 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.222862 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.222881 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:50Z","lastTransitionTime":"2026-01-26T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.325937 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.326052 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.326075 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.326105 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.326125 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:50Z","lastTransitionTime":"2026-01-26T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.430119 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.430186 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.430198 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.430219 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.430232 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:50Z","lastTransitionTime":"2026-01-26T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.533339 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.533408 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.533427 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.533454 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.533478 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:50Z","lastTransitionTime":"2026-01-26T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.636907 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.637275 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.637352 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.637421 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.637477 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:50Z","lastTransitionTime":"2026-01-26T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.740910 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.741290 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.741414 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.741504 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.741574 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:50Z","lastTransitionTime":"2026-01-26T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.844873 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.844928 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.844951 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.844990 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.845002 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:50Z","lastTransitionTime":"2026-01-26T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.948192 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.948239 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.948249 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.948267 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:50 crc kubenswrapper[4872]: I0126 09:08:50.948279 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:50Z","lastTransitionTime":"2026-01-26T09:08:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.051057 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.051095 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.051104 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.051120 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.051130 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:51Z","lastTransitionTime":"2026-01-26T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.154186 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.154260 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.154279 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.154314 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.154333 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:51Z","lastTransitionTime":"2026-01-26T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.157922 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 15:29:30.533171397 +0000 UTC Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.184008 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:51 crc kubenswrapper[4872]: E0126 09:08:51.184192 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.184297 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:51 crc kubenswrapper[4872]: E0126 09:08:51.184452 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.184464 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:51 crc kubenswrapper[4872]: E0126 09:08:51.184551 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.184705 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:51 crc kubenswrapper[4872]: E0126 09:08:51.184768 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.257944 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.258027 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.258047 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.258079 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.258102 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:51Z","lastTransitionTime":"2026-01-26T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.360941 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.361021 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.361032 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.361053 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.361073 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:51Z","lastTransitionTime":"2026-01-26T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.463982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.464030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.464044 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.464060 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.464074 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:51Z","lastTransitionTime":"2026-01-26T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.567024 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.567075 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.567090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.567111 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.567125 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:51Z","lastTransitionTime":"2026-01-26T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.670062 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.670113 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.670134 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.670159 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.670176 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:51Z","lastTransitionTime":"2026-01-26T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.772655 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.772711 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.772730 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.772753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.772769 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:51Z","lastTransitionTime":"2026-01-26T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.875202 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.875289 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.875304 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.875336 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.875351 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:51Z","lastTransitionTime":"2026-01-26T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.977442 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.977509 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.977532 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.977552 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:51 crc kubenswrapper[4872]: I0126 09:08:51.977569 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:51Z","lastTransitionTime":"2026-01-26T09:08:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.080381 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.080424 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.080436 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.080453 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.080465 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:52Z","lastTransitionTime":"2026-01-26T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.158455 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 10:07:05.704298511 +0000 UTC Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.182588 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.182619 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.182628 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.182640 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.182650 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:52Z","lastTransitionTime":"2026-01-26T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.285509 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.285580 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.285595 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.285619 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.285632 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:52Z","lastTransitionTime":"2026-01-26T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.387972 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.388018 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.388028 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.388043 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.388057 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:52Z","lastTransitionTime":"2026-01-26T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.491368 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.491413 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.491424 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.491441 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.491452 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:52Z","lastTransitionTime":"2026-01-26T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.594328 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.594368 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.594379 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.594395 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.594406 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:52Z","lastTransitionTime":"2026-01-26T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.697144 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.697189 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.697229 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.697251 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.697267 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:52Z","lastTransitionTime":"2026-01-26T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.800889 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.800986 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.801004 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.801068 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.801087 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:52Z","lastTransitionTime":"2026-01-26T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.903580 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.903648 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.903660 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.903680 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.903696 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:52Z","lastTransitionTime":"2026-01-26T09:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:52 crc kubenswrapper[4872]: I0126 09:08:52.986390 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:52 crc kubenswrapper[4872]: E0126 09:08:52.986605 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:52 crc kubenswrapper[4872]: E0126 09:08:52.986708 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs podName:8f34a04d-444c-4640-9b51-51850a6317e6 nodeName:}" failed. No retries permitted until 2026-01-26 09:09:24.986675921 +0000 UTC m=+98.295515762 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs") pod "network-metrics-daemon-krd2b" (UID: "8f34a04d-444c-4640-9b51-51850a6317e6") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.006648 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.006686 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.006695 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.006712 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.006723 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:53Z","lastTransitionTime":"2026-01-26T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.110163 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.110216 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.110226 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.110243 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.110255 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:53Z","lastTransitionTime":"2026-01-26T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.141599 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" probeResult="failure" output="" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.159125 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 21:43:55.745864282 +0000 UTC Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.184008 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.184035 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.184126 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.184102 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:53 crc kubenswrapper[4872]: E0126 09:08:53.184182 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:53 crc kubenswrapper[4872]: E0126 09:08:53.184311 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:53 crc kubenswrapper[4872]: E0126 09:08:53.184755 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:53 crc kubenswrapper[4872]: E0126 09:08:53.184904 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.212939 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.212992 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.213008 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.213030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.213043 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:53Z","lastTransitionTime":"2026-01-26T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.315660 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.315707 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.315719 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.315739 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.315753 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:53Z","lastTransitionTime":"2026-01-26T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.418189 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.418284 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.418351 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.418377 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.418437 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:53Z","lastTransitionTime":"2026-01-26T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.522182 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.522241 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.522259 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.522283 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.522299 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:53Z","lastTransitionTime":"2026-01-26T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.631755 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.631792 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.631862 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.631887 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.631900 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:53Z","lastTransitionTime":"2026-01-26T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.735118 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.735158 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.735197 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.735215 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.735226 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:53Z","lastTransitionTime":"2026-01-26T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.838015 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.838064 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.838076 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.838092 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.838104 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:53Z","lastTransitionTime":"2026-01-26T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.940868 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.940914 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.940929 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.940946 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:53 crc kubenswrapper[4872]: I0126 09:08:53.940957 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:53Z","lastTransitionTime":"2026-01-26T09:08:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.043157 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.043190 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.043198 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.043213 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.043222 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:54Z","lastTransitionTime":"2026-01-26T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.146340 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.146391 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.146402 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.146418 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.146430 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:54Z","lastTransitionTime":"2026-01-26T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.160068 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 00:27:08.195568585 +0000 UTC Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.249285 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.249354 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.249367 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.249413 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.249427 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:54Z","lastTransitionTime":"2026-01-26T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.352256 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.352287 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.352298 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.352313 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.352324 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:54Z","lastTransitionTime":"2026-01-26T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.455775 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.455844 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.455863 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.455879 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.455889 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:54Z","lastTransitionTime":"2026-01-26T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.559100 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.559201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.559234 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.559268 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.559291 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:54Z","lastTransitionTime":"2026-01-26T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.662071 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.662133 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.662151 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.662174 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.662191 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:54Z","lastTransitionTime":"2026-01-26T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.685161 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-swcf5_3579107d-a172-4905-a491-6dd2b5a304a5/kube-multus/0.log" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.685245 4872 generic.go:334] "Generic (PLEG): container finished" podID="3579107d-a172-4905-a491-6dd2b5a304a5" containerID="27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924" exitCode=1 Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.685290 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-swcf5" event={"ID":"3579107d-a172-4905-a491-6dd2b5a304a5","Type":"ContainerDied","Data":"27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924"} Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.686012 4872 scope.go:117] "RemoveContainer" containerID="27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.705656 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.721748 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.740073 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.756756 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.765305 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.765357 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.765376 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.765401 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.765418 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:54Z","lastTransitionTime":"2026-01-26T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.774026 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.787917 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.804051 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.818645 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.833703 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13bb0fc1-b822-4651-ab5f-b095c86e936b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1754e95c65f24f7086f23e38595f8d3b6acb9baaaed809afee607eb5a13d06bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178c8017e99cb8783654f3249f3499f6e61c7cabfbedf1bd078e3f345b3e7cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3c7b1aa5169eb070e390f0b0e5f06149b73145a6a041e0c3e6ba46857f77814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.857782 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.868466 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.868514 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.868527 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.868546 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.868559 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:54Z","lastTransitionTime":"2026-01-26T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.873517 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.894301 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.915816 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.935091 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.959168 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.973266 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.973607 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.973919 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.974252 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.974452 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:54Z","lastTransitionTime":"2026-01-26T09:08:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:54 crc kubenswrapper[4872]: I0126 09:08:54.979384 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.002240 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:54Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.023378 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:54Z\\\",\\\"message\\\":\\\"2026-01-26T09:08:09+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_04bb5e40-547e-4db5-9678-a80684512fdc\\\\n2026-01-26T09:08:09+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_04bb5e40-547e-4db5-9678-a80684512fdc to /host/opt/cni/bin/\\\\n2026-01-26T09:08:09Z [verbose] multus-daemon started\\\\n2026-01-26T09:08:09Z [verbose] Readiness Indicator file check\\\\n2026-01-26T09:08:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.077978 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.078251 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.078316 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.078396 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.078470 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:55Z","lastTransitionTime":"2026-01-26T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.161066 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 15:41:40.655845159 +0000 UTC Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.182048 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.182088 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.182097 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.182113 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.182123 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:55Z","lastTransitionTime":"2026-01-26T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.182976 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.183098 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:55 crc kubenswrapper[4872]: E0126 09:08:55.183195 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:55 crc kubenswrapper[4872]: E0126 09:08:55.183323 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.183510 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:55 crc kubenswrapper[4872]: E0126 09:08:55.183636 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.183883 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:55 crc kubenswrapper[4872]: E0126 09:08:55.184252 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.285762 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.285936 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.285964 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.285999 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.286574 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:55Z","lastTransitionTime":"2026-01-26T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.399912 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.400327 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.400516 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.400684 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.401021 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:55Z","lastTransitionTime":"2026-01-26T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.503356 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.503413 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.503426 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.503447 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.503461 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:55Z","lastTransitionTime":"2026-01-26T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.605609 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.605670 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.605680 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.605695 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.605706 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:55Z","lastTransitionTime":"2026-01-26T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.690995 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-swcf5_3579107d-a172-4905-a491-6dd2b5a304a5/kube-multus/0.log" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.691068 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-swcf5" event={"ID":"3579107d-a172-4905-a491-6dd2b5a304a5","Type":"ContainerStarted","Data":"4960fa5dfa153921fd4c32dffe9662db8304b2a8242ef2b570ccc174c76279d8"} Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.706070 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.707366 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.707404 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.707415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.707432 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.707446 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:55Z","lastTransitionTime":"2026-01-26T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.720603 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.732975 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.744918 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.760037 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.773281 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.788837 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.800913 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.810492 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.810712 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.810903 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.811053 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.811182 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:55Z","lastTransitionTime":"2026-01-26T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.813969 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13bb0fc1-b822-4651-ab5f-b095c86e936b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1754e95c65f24f7086f23e38595f8d3b6acb9baaaed809afee607eb5a13d06bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178c8017e99cb8783654f3249f3499f6e61c7cabfbedf1bd078e3f345b3e7cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3c7b1aa5169eb070e390f0b0e5f06149b73145a6a041e0c3e6ba46857f77814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.836341 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.845464 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.859439 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.875903 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.890200 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.913642 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.913919 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.914027 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.914158 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.914294 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:55Z","lastTransitionTime":"2026-01-26T09:08:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.942636 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.958197 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.976867 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:55 crc kubenswrapper[4872]: I0126 09:08:55.990136 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4960fa5dfa153921fd4c32dffe9662db8304b2a8242ef2b570ccc174c76279d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:54Z\\\",\\\"message\\\":\\\"2026-01-26T09:08:09+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_04bb5e40-547e-4db5-9678-a80684512fdc\\\\n2026-01-26T09:08:09+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_04bb5e40-547e-4db5-9678-a80684512fdc to /host/opt/cni/bin/\\\\n2026-01-26T09:08:09Z [verbose] multus-daemon started\\\\n2026-01-26T09:08:09Z [verbose] Readiness Indicator file check\\\\n2026-01-26T09:08:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:55Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.018050 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.018109 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.018122 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.018142 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.018156 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:56Z","lastTransitionTime":"2026-01-26T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.121666 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.121733 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.121753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.121781 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.121827 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:56Z","lastTransitionTime":"2026-01-26T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.162140 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 09:19:58.467974178 +0000 UTC Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.196209 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.224473 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.224534 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.224550 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.224575 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.224588 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:56Z","lastTransitionTime":"2026-01-26T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.327068 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.327133 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.327146 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.327176 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.327192 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:56Z","lastTransitionTime":"2026-01-26T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.430081 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.430122 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.430132 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.430148 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.430160 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:56Z","lastTransitionTime":"2026-01-26T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.532534 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.532586 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.532598 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.532616 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.532629 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:56Z","lastTransitionTime":"2026-01-26T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.634605 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.634652 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.634664 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.634682 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.634695 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:56Z","lastTransitionTime":"2026-01-26T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.737945 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.738004 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.738021 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.738047 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.738066 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:56Z","lastTransitionTime":"2026-01-26T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.840680 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.840717 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.840729 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.840743 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.840752 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:56Z","lastTransitionTime":"2026-01-26T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.944352 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.944415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.944433 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.944457 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:56 crc kubenswrapper[4872]: I0126 09:08:56.944474 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:56Z","lastTransitionTime":"2026-01-26T09:08:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.046825 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.046857 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.046866 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.046883 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.046893 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:57Z","lastTransitionTime":"2026-01-26T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.149053 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.149102 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.149115 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.149134 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.149146 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:57Z","lastTransitionTime":"2026-01-26T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.163382 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 18:18:31.152199966 +0000 UTC Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.187959 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:57 crc kubenswrapper[4872]: E0126 09:08:57.188253 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.188880 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.188941 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:57 crc kubenswrapper[4872]: E0126 09:08:57.189031 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:57 crc kubenswrapper[4872]: E0126 09:08:57.189123 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.189244 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:57 crc kubenswrapper[4872]: E0126 09:08:57.189425 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.200563 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.213876 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13bb0fc1-b822-4651-ab5f-b095c86e936b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1754e95c65f24f7086f23e38595f8d3b6acb9baaaed809afee607eb5a13d06bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178c8017e99cb8783654f3249f3499f6e61c7cabfbedf1bd078e3f345b3e7cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3c7b1aa5169eb070e390f0b0e5f06149b73145a6a041e0c3e6ba46857f77814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.234246 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.251639 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.252373 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.252468 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.252533 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.252605 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.252685 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:57Z","lastTransitionTime":"2026-01-26T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.266852 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.282919 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.294715 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.306517 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4960fa5dfa153921fd4c32dffe9662db8304b2a8242ef2b570ccc174c76279d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:54Z\\\",\\\"message\\\":\\\"2026-01-26T09:08:09+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_04bb5e40-547e-4db5-9678-a80684512fdc\\\\n2026-01-26T09:08:09+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_04bb5e40-547e-4db5-9678-a80684512fdc to /host/opt/cni/bin/\\\\n2026-01-26T09:08:09Z [verbose] multus-daemon started\\\\n2026-01-26T09:08:09Z [verbose] Readiness Indicator file check\\\\n2026-01-26T09:08:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.321045 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.339633 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.350562 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.354643 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.354666 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.354674 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.354688 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.354697 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:57Z","lastTransitionTime":"2026-01-26T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.366940 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.382303 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.394766 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.407635 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.421321 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.433341 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a28903ba-307e-40de-9adf-9ff4cb83731d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ce2232b80febe7ce3f28e689b9ac4b270d6c623038a1eb0e65a336d4e9aa34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc1cf165c53ff3c104af93fdeff6c89169e4b6acf7ea669c4a30468359610483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc1cf165c53ff3c104af93fdeff6c89169e4b6acf7ea669c4a30468359610483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.448331 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.457673 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.457709 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.457719 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.457735 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.457749 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:57Z","lastTransitionTime":"2026-01-26T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.460749 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:57Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.559879 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.559927 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.559945 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.559968 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.559988 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:57Z","lastTransitionTime":"2026-01-26T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.662330 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.662390 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.662407 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.662429 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.662446 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:57Z","lastTransitionTime":"2026-01-26T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.765404 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.765437 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.765448 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.765463 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.765473 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:57Z","lastTransitionTime":"2026-01-26T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.867699 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.867730 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.867741 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.867758 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.867768 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:57Z","lastTransitionTime":"2026-01-26T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.971072 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.971109 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.971118 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.971134 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:57 crc kubenswrapper[4872]: I0126 09:08:57.971143 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:57Z","lastTransitionTime":"2026-01-26T09:08:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.074118 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.074168 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.074181 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.074201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.074217 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:58Z","lastTransitionTime":"2026-01-26T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.164447 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-20 12:32:08.897615095 +0000 UTC Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.176233 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.176296 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.176310 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.176329 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.176348 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:58Z","lastTransitionTime":"2026-01-26T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.278873 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.278953 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.278974 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.278996 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.279013 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:58Z","lastTransitionTime":"2026-01-26T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.381432 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.381481 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.381495 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.381513 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.381525 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:58Z","lastTransitionTime":"2026-01-26T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.483884 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.483922 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.483937 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.483952 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.483965 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:58Z","lastTransitionTime":"2026-01-26T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.586891 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.586942 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.586954 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.586973 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.586985 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:58Z","lastTransitionTime":"2026-01-26T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.691554 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.691587 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.691597 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.691613 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.691623 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:58Z","lastTransitionTime":"2026-01-26T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.702895 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovnkube-controller/2.log" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.703659 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovnkube-controller/1.log" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.707518 4872 generic.go:334] "Generic (PLEG): container finished" podID="e2577331-1886-4abb-896b-8d8a112e40d8" containerID="e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621" exitCode=1 Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.707565 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621"} Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.707601 4872 scope.go:117] "RemoveContainer" containerID="40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.708336 4872 scope.go:117] "RemoveContainer" containerID="e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621" Jan 26 09:08:58 crc kubenswrapper[4872]: E0126 09:08:58.708526 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6p5bj_openshift-ovn-kubernetes(e2577331-1886-4abb-896b-8d8a112e40d8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.730876 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.754339 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.770081 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4960fa5dfa153921fd4c32dffe9662db8304b2a8242ef2b570ccc174c76279d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:54Z\\\",\\\"message\\\":\\\"2026-01-26T09:08:09+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_04bb5e40-547e-4db5-9678-a80684512fdc\\\\n2026-01-26T09:08:09+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_04bb5e40-547e-4db5-9678-a80684512fdc to /host/opt/cni/bin/\\\\n2026-01-26T09:08:09Z [verbose] multus-daemon started\\\\n2026-01-26T09:08:09Z [verbose] Readiness Indicator file check\\\\n2026-01-26T09:08:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.784168 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.794039 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.794088 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.794100 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.794119 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.794133 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:58Z","lastTransitionTime":"2026-01-26T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.796254 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.809735 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.820766 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.830601 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.839980 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a28903ba-307e-40de-9adf-9ff4cb83731d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ce2232b80febe7ce3f28e689b9ac4b270d6c623038a1eb0e65a336d4e9aa34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc1cf165c53ff3c104af93fdeff6c89169e4b6acf7ea669c4a30468359610483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc1cf165c53ff3c104af93fdeff6c89169e4b6acf7ea669c4a30468359610483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.850636 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.861932 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.872777 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.883998 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13bb0fc1-b822-4651-ab5f-b095c86e936b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1754e95c65f24f7086f23e38595f8d3b6acb9baaaed809afee607eb5a13d06bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178c8017e99cb8783654f3249f3499f6e61c7cabfbedf1bd078e3f345b3e7cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3c7b1aa5169eb070e390f0b0e5f06149b73145a6a041e0c3e6ba46857f77814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.896562 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.896620 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.896635 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.896650 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.897020 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:58Z","lastTransitionTime":"2026-01-26T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.907912 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:58Z\\\",\\\"message\\\":\\\"ager: error found while processing openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk: failed to check if pod openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk is in primary UDN: could not find OVN pod annotation in map[]\\\\nI0126 09:08:57.995458 6522 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-dns-operator/dns-operator-744455d44c-w5lkb: failed to check if pod openshift-dns-operator/dns-operator-744455d44c-w5lkb is in primary UDN: could not find OVN pod annotation in map[]\\\\nI0126 09:08:57.995521 6522 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7: failed to check if pod openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7 is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:restricted-v2 openshift.io/scc:restricted-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nE0126 09:08:58.086544 6522 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI0126 09:08:58.087737 6522 ovnkube.go:599] Stopped ovnkube\\\\nI0126 09:08:58.087837 6522 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.918328 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.931330 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.945009 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.957500 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.968386 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:58Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:58 crc kubenswrapper[4872]: I0126 09:08:58.999850 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:58.999936 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:58.999957 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.000394 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.000633 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:58Z","lastTransitionTime":"2026-01-26T09:08:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.103520 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.103566 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.103623 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.103643 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.103655 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.164862 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 23:36:04.317401328 +0000 UTC Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.183384 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.183460 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.183404 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.183396 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:08:59 crc kubenswrapper[4872]: E0126 09:08:59.183589 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:08:59 crc kubenswrapper[4872]: E0126 09:08:59.183738 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:08:59 crc kubenswrapper[4872]: E0126 09:08:59.183905 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:08:59 crc kubenswrapper[4872]: E0126 09:08:59.184053 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.206445 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.206489 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.206501 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.206518 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.206529 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.309051 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.309098 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.309110 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.309129 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.309141 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.412327 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.412394 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.412408 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.412429 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.412442 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.514749 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.514814 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.514826 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.514845 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.514857 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.617315 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.617354 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.617365 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.617384 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.617394 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.634838 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.634878 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.634893 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.634912 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.634928 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: E0126 09:08:59.648081 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:59Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.651634 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.651667 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.651680 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.651717 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.651729 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: E0126 09:08:59.665025 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:59Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.668307 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.668346 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.668358 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.668374 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.668386 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: E0126 09:08:59.681318 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:59Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.685502 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.685548 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.685578 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.685600 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.685614 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: E0126 09:08:59.705536 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:59Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.711623 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.711656 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.711669 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.711686 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.711699 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.714914 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovnkube-controller/2.log" Jan 26 09:08:59 crc kubenswrapper[4872]: E0126 09:08:59.725000 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:08:59Z is after 2025-08-24T17:21:41Z" Jan 26 09:08:59 crc kubenswrapper[4872]: E0126 09:08:59.725107 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.726982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.727013 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.727025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.727038 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.727049 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.829620 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.829658 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.829667 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.829680 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.829689 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.932553 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.932609 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.932630 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.932659 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:08:59 crc kubenswrapper[4872]: I0126 09:08:59.932679 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:08:59Z","lastTransitionTime":"2026-01-26T09:08:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.036127 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.036232 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.036259 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.036289 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.036311 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:00Z","lastTransitionTime":"2026-01-26T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.139033 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.139082 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.139096 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.139114 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.139126 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:00Z","lastTransitionTime":"2026-01-26T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.165953 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 19:08:55.503606689 +0000 UTC Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.241428 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.241463 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.241473 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.241487 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.241497 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:00Z","lastTransitionTime":"2026-01-26T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.344688 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.344750 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.344761 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.344792 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.344820 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:00Z","lastTransitionTime":"2026-01-26T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.447235 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.447279 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.447294 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.447313 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.447325 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:00Z","lastTransitionTime":"2026-01-26T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.549866 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.549932 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.549949 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.549981 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.549998 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:00Z","lastTransitionTime":"2026-01-26T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.652771 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.653007 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.653051 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.653080 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.653097 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:00Z","lastTransitionTime":"2026-01-26T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.757615 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.757672 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.757682 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.757700 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.757711 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:00Z","lastTransitionTime":"2026-01-26T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.861244 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.861305 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.861322 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.861347 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.861365 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:00Z","lastTransitionTime":"2026-01-26T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.964042 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.964124 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.964149 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.964180 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:00 crc kubenswrapper[4872]: I0126 09:09:00.964202 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:00Z","lastTransitionTime":"2026-01-26T09:09:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.067326 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.067406 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.067534 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.067564 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.067583 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:01Z","lastTransitionTime":"2026-01-26T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.166908 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-06 10:08:02.590357046 +0000 UTC Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.170491 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.170539 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.170551 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.170573 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.170592 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:01Z","lastTransitionTime":"2026-01-26T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.183165 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:01 crc kubenswrapper[4872]: E0126 09:09:01.183429 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.183692 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:01 crc kubenswrapper[4872]: E0126 09:09:01.183777 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.183967 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:01 crc kubenswrapper[4872]: E0126 09:09:01.184028 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.184263 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:01 crc kubenswrapper[4872]: E0126 09:09:01.184339 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.274055 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.274135 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.274153 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.274181 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.274204 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:01Z","lastTransitionTime":"2026-01-26T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.376995 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.377030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.377039 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.377052 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.377061 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:01Z","lastTransitionTime":"2026-01-26T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.480369 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.480418 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.480435 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.480456 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.480473 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:01Z","lastTransitionTime":"2026-01-26T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.583025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.583078 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.583093 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.583116 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.583135 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:01Z","lastTransitionTime":"2026-01-26T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.686519 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.686555 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.686566 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.686582 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.686592 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:01Z","lastTransitionTime":"2026-01-26T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.789674 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.790079 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.790099 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.790123 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.790137 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:01Z","lastTransitionTime":"2026-01-26T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.893541 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.893603 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.893617 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.893638 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.893652 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:01Z","lastTransitionTime":"2026-01-26T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.997253 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.997329 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.997352 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.997380 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:01 crc kubenswrapper[4872]: I0126 09:09:01.997399 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:01Z","lastTransitionTime":"2026-01-26T09:09:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.101283 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.101375 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.101393 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.101417 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.101430 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:02Z","lastTransitionTime":"2026-01-26T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.167306 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 13:49:47.696458982 +0000 UTC Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.204634 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.204731 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.204756 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.204787 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.204859 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:02Z","lastTransitionTime":"2026-01-26T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.308070 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.308140 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.308153 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.308175 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.308189 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:02Z","lastTransitionTime":"2026-01-26T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.411875 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.411918 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.411931 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.411955 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.411969 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:02Z","lastTransitionTime":"2026-01-26T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.515415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.515455 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.515465 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.515487 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.515497 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:02Z","lastTransitionTime":"2026-01-26T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.619181 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.619255 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.619271 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.619300 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.619319 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:02Z","lastTransitionTime":"2026-01-26T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.723047 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.723118 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.723132 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.723171 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.723216 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:02Z","lastTransitionTime":"2026-01-26T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.827135 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.827202 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.827213 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.827235 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.827249 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:02Z","lastTransitionTime":"2026-01-26T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.930755 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.930893 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.930916 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.930943 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:02 crc kubenswrapper[4872]: I0126 09:09:02.931041 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:02Z","lastTransitionTime":"2026-01-26T09:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.034225 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.034280 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.034290 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.034311 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.034323 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:03Z","lastTransitionTime":"2026-01-26T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.138287 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.138363 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.138373 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.138392 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.138405 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:03Z","lastTransitionTime":"2026-01-26T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.168467 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-16 01:04:10.037824143 +0000 UTC Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.182956 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:03 crc kubenswrapper[4872]: E0126 09:09:03.183138 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.183371 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:03 crc kubenswrapper[4872]: E0126 09:09:03.183457 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.183377 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:03 crc kubenswrapper[4872]: E0126 09:09:03.183533 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.183482 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:03 crc kubenswrapper[4872]: E0126 09:09:03.183591 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.241298 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.241369 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.241388 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.241409 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.241422 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:03Z","lastTransitionTime":"2026-01-26T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.344535 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.344598 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.344614 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.344640 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.344657 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:03Z","lastTransitionTime":"2026-01-26T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.447515 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.447567 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.447576 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.447594 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.447606 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:03Z","lastTransitionTime":"2026-01-26T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.550568 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.550616 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.550625 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.550641 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.550652 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:03Z","lastTransitionTime":"2026-01-26T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.652896 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.652955 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.652967 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.652987 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.653057 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:03Z","lastTransitionTime":"2026-01-26T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.756693 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.757035 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.757055 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.757078 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.757090 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:03Z","lastTransitionTime":"2026-01-26T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.860135 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.860192 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.860201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.860222 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.860232 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:03Z","lastTransitionTime":"2026-01-26T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.963363 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.963418 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.963430 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.963452 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:03 crc kubenswrapper[4872]: I0126 09:09:03.963466 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:03Z","lastTransitionTime":"2026-01-26T09:09:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.066412 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.066496 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.066521 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.066554 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.066580 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:04Z","lastTransitionTime":"2026-01-26T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.169161 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 03:45:06.80645845 +0000 UTC Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.169758 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.169811 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.169821 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.169840 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.169851 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:04Z","lastTransitionTime":"2026-01-26T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.272925 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.272985 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.272998 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.273020 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.273033 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:04Z","lastTransitionTime":"2026-01-26T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.376517 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.376561 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.376573 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.376592 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.376604 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:04Z","lastTransitionTime":"2026-01-26T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.480447 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.480495 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.480505 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.480521 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.480531 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:04Z","lastTransitionTime":"2026-01-26T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.582775 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.582859 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.582868 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.582882 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.582920 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:04Z","lastTransitionTime":"2026-01-26T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.685352 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.685397 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.685408 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.685445 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.685478 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:04Z","lastTransitionTime":"2026-01-26T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.788404 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.788479 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.788501 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.788530 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.788558 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:04Z","lastTransitionTime":"2026-01-26T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.892204 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.892288 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.892321 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.892355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.892377 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:04Z","lastTransitionTime":"2026-01-26T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.996013 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.996076 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.996090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.996112 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:04 crc kubenswrapper[4872]: I0126 09:09:04.996128 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:04Z","lastTransitionTime":"2026-01-26T09:09:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.098686 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.098745 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.098759 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.098779 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.098792 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:05Z","lastTransitionTime":"2026-01-26T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.169402 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-14 15:55:53.572127439 +0000 UTC Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.183142 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:05 crc kubenswrapper[4872]: E0126 09:09:05.183315 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.183163 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.183394 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:05 crc kubenswrapper[4872]: E0126 09:09:05.183481 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.183142 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:05 crc kubenswrapper[4872]: E0126 09:09:05.183579 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:05 crc kubenswrapper[4872]: E0126 09:09:05.183635 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.201876 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.201928 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.201940 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.201959 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.201973 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:05Z","lastTransitionTime":"2026-01-26T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.305420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.305499 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.305509 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.305530 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.305544 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:05Z","lastTransitionTime":"2026-01-26T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.408724 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.408788 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.408813 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.408832 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.408846 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:05Z","lastTransitionTime":"2026-01-26T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.512336 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.512386 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.512421 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.512445 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.512459 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:05Z","lastTransitionTime":"2026-01-26T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.649396 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.649462 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.649476 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.649496 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.649509 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:05Z","lastTransitionTime":"2026-01-26T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.752054 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.752149 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.752180 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.752217 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.752241 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:05Z","lastTransitionTime":"2026-01-26T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.855547 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.855600 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.855609 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.855628 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.855640 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:05Z","lastTransitionTime":"2026-01-26T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.959217 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.959326 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.959346 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.959377 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:05 crc kubenswrapper[4872]: I0126 09:09:05.959408 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:05Z","lastTransitionTime":"2026-01-26T09:09:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.062241 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.062315 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.062328 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.062350 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.062366 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:06Z","lastTransitionTime":"2026-01-26T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.165741 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.165840 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.165866 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.165907 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.165932 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:06Z","lastTransitionTime":"2026-01-26T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.169865 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-16 04:04:48.324063398 +0000 UTC Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.268598 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.268633 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.268643 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.268662 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.268675 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:06Z","lastTransitionTime":"2026-01-26T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.372400 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.372764 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.372883 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.373053 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.373200 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:06Z","lastTransitionTime":"2026-01-26T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.477182 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.477238 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.477256 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.477284 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.477303 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:06Z","lastTransitionTime":"2026-01-26T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.579951 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.580266 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.580326 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.580404 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.580463 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:06Z","lastTransitionTime":"2026-01-26T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.683840 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.683874 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.683885 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.683901 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.683912 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:06Z","lastTransitionTime":"2026-01-26T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.786752 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.786854 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.786866 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.786882 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.786893 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:06Z","lastTransitionTime":"2026-01-26T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.889864 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.889926 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.889939 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.889956 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.889968 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:06Z","lastTransitionTime":"2026-01-26T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.993427 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.993475 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.993485 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.993503 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:06 crc kubenswrapper[4872]: I0126 09:09:06.993516 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:06Z","lastTransitionTime":"2026-01-26T09:09:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.095945 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.096006 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.096018 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.096037 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.096050 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:07Z","lastTransitionTime":"2026-01-26T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.170941 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 03:14:51.609064688 +0000 UTC Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.183482 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.183568 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:07 crc kubenswrapper[4872]: E0126 09:09:07.184533 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.184881 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.184987 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:07 crc kubenswrapper[4872]: E0126 09:09:07.185093 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:07 crc kubenswrapper[4872]: E0126 09:09:07.185413 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:07 crc kubenswrapper[4872]: E0126 09:09:07.185626 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.199160 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.199222 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.199235 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.199256 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.199268 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:07Z","lastTransitionTime":"2026-01-26T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.204104 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.225407 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.241445 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.260117 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.282574 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.300291 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a28903ba-307e-40de-9adf-9ff4cb83731d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ce2232b80febe7ce3f28e689b9ac4b270d6c623038a1eb0e65a336d4e9aa34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc1cf165c53ff3c104af93fdeff6c89169e4b6acf7ea669c4a30468359610483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc1cf165c53ff3c104af93fdeff6c89169e4b6acf7ea669c4a30468359610483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.303506 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.303537 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.303549 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.303571 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.303585 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:07Z","lastTransitionTime":"2026-01-26T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.317192 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.330912 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.345461 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.358672 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13bb0fc1-b822-4651-ab5f-b095c86e936b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1754e95c65f24f7086f23e38595f8d3b6acb9baaaed809afee607eb5a13d06bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178c8017e99cb8783654f3249f3499f6e61c7cabfbedf1bd078e3f345b3e7cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3c7b1aa5169eb070e390f0b0e5f06149b73145a6a041e0c3e6ba46857f77814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.378074 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40cef623271201999439c8997948b96ed9680eda3ed072cdb3db615b901cf6a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:18Z\\\",\\\"message\\\":\\\"efault, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"20da2226-531c-4179-9810-aa4026995ca3\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-marketplace/certified-operators_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-marketplace/certified-operators\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.214\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0126 09:08:18.321348 6303 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0126 09:08:18.321437 6303 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:17Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:58Z\\\",\\\"message\\\":\\\"ager: error found while processing openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk: failed to check if pod openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk is in primary UDN: could not find OVN pod annotation in map[]\\\\nI0126 09:08:57.995458 6522 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-dns-operator/dns-operator-744455d44c-w5lkb: failed to check if pod openshift-dns-operator/dns-operator-744455d44c-w5lkb is in primary UDN: could not find OVN pod annotation in map[]\\\\nI0126 09:08:57.995521 6522 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7: failed to check if pod openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7 is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:restricted-v2 openshift.io/scc:restricted-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nE0126 09:08:58.086544 6522 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI0126 09:08:58.087737 6522 ovnkube.go:599] Stopped ovnkube\\\\nI0126 09:08:58.087837 6522 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.388565 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.405773 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.407543 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.407653 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.407721 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.407791 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.407891 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:07Z","lastTransitionTime":"2026-01-26T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.423683 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.438681 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.451639 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.467577 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.486066 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.502199 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4960fa5dfa153921fd4c32dffe9662db8304b2a8242ef2b570ccc174c76279d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:54Z\\\",\\\"message\\\":\\\"2026-01-26T09:08:09+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_04bb5e40-547e-4db5-9678-a80684512fdc\\\\n2026-01-26T09:08:09+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_04bb5e40-547e-4db5-9678-a80684512fdc to /host/opt/cni/bin/\\\\n2026-01-26T09:08:09Z [verbose] multus-daemon started\\\\n2026-01-26T09:08:09Z [verbose] Readiness Indicator file check\\\\n2026-01-26T09:08:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:07Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.510308 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.510353 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.510365 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.510383 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.510397 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:07Z","lastTransitionTime":"2026-01-26T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.614378 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.614410 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.614419 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.614433 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.614442 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:07Z","lastTransitionTime":"2026-01-26T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.717549 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.717598 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.717613 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.717632 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.717647 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:07Z","lastTransitionTime":"2026-01-26T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.822024 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.822087 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.822103 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.822127 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.822145 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:07Z","lastTransitionTime":"2026-01-26T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.925038 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.925467 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.925566 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.925690 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:07 crc kubenswrapper[4872]: I0126 09:09:07.925788 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:07Z","lastTransitionTime":"2026-01-26T09:09:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.028279 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.028335 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.028345 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.028364 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.028375 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:08Z","lastTransitionTime":"2026-01-26T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.131993 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.132052 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.132063 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.132081 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.132092 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:08Z","lastTransitionTime":"2026-01-26T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.171435 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 08:33:57.812979777 +0000 UTC Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.234885 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.234937 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.234946 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.234964 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.234975 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:08Z","lastTransitionTime":"2026-01-26T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.337764 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.337843 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.337855 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.337870 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.337880 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:08Z","lastTransitionTime":"2026-01-26T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.441419 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.441474 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.441484 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.441505 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.441515 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:08Z","lastTransitionTime":"2026-01-26T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.544714 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.544774 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.544787 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.544845 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.544861 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:08Z","lastTransitionTime":"2026-01-26T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.649370 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.649427 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.649442 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.649464 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.649479 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:08Z","lastTransitionTime":"2026-01-26T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.753589 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.753640 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.753650 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.753672 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.753685 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:08Z","lastTransitionTime":"2026-01-26T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.856704 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.856773 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.856791 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.856842 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.856860 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:08Z","lastTransitionTime":"2026-01-26T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.959899 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.959956 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.960005 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.960030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.960043 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:08Z","lastTransitionTime":"2026-01-26T09:09:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:08 crc kubenswrapper[4872]: I0126 09:09:08.987545 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:08 crc kubenswrapper[4872]: E0126 09:09:08.987735 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:10:12.987681972 +0000 UTC m=+146.296521933 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.062524 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.062584 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.062596 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.062617 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.062627 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:09Z","lastTransitionTime":"2026-01-26T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.088985 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.089034 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.089077 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.089096 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.089207 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.089247 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.089261 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.089219 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.089321 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:10:13.089297572 +0000 UTC m=+146.398137373 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.089322 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.089345 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.089248 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.089376 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-01-26 09:10:13.089367714 +0000 UTC m=+146.398207615 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.089275 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.089405 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-01-26 09:10:13.089394655 +0000 UTC m=+146.398234456 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.089418 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-01-26 09:10:13.089410775 +0000 UTC m=+146.398250676 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.165987 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.166063 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.166083 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.166113 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.166136 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:09Z","lastTransitionTime":"2026-01-26T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.173242 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 09:04:33.653237235 +0000 UTC Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.183795 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.183907 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.183965 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.184090 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.184250 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.184468 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.184588 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:09 crc kubenswrapper[4872]: E0126 09:09:09.184710 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.269395 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.269845 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.270037 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.270391 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.270612 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:09Z","lastTransitionTime":"2026-01-26T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.374490 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.374931 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.375138 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.375321 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.375585 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:09Z","lastTransitionTime":"2026-01-26T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.477846 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.477903 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.477915 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.477936 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.477949 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:09Z","lastTransitionTime":"2026-01-26T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.580456 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.580506 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.580514 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.580534 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.580548 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:09Z","lastTransitionTime":"2026-01-26T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.683225 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.683274 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.683294 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.683316 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.683330 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:09Z","lastTransitionTime":"2026-01-26T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.785986 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.786037 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.786051 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.786071 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.786084 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:09Z","lastTransitionTime":"2026-01-26T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.889364 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.889430 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.889447 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.889477 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.889491 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:09Z","lastTransitionTime":"2026-01-26T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.992658 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.992698 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.992708 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.992724 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:09 crc kubenswrapper[4872]: I0126 09:09:09.992733 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:09Z","lastTransitionTime":"2026-01-26T09:09:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.072218 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.072278 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.072289 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.072309 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.072321 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: E0126 09:09:10.092511 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.098351 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.098495 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.098517 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.098567 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.098600 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: E0126 09:09:10.121735 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.127311 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.127359 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.127371 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.127390 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.127402 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: E0126 09:09:10.143974 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.149631 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.149686 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.149702 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.149724 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.149739 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: E0126 09:09:10.164560 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.168745 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.168833 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.168853 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.168877 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.168893 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.174984 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 12:26:27.831124237 +0000 UTC Jan 26 09:09:10 crc kubenswrapper[4872]: E0126 09:09:10.184785 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-01-26T09:09:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"99e27e58-03f3-4e41-9557-1ea806d6558d\\\",\\\"systemUUID\\\":\\\"9981fead-c93d-4b79-8f75-eea6ea7db214\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:10Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:10 crc kubenswrapper[4872]: E0126 09:09:10.185053 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.187598 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.187656 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.187675 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.187702 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.187719 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.291443 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.291508 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.291521 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.291541 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.291560 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.394317 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.394648 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.394717 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.394812 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.394886 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.497875 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.497959 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.497972 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.497997 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.498016 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.600700 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.602014 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.602086 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.602118 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.602136 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.706334 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.706393 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.706414 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.706443 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.706463 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.808778 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.809034 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.809138 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.809222 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.809290 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.912436 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.913005 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.913030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.913064 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:10 crc kubenswrapper[4872]: I0126 09:09:10.913083 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:10Z","lastTransitionTime":"2026-01-26T09:09:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.015604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.016197 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.016279 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.016372 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.016444 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:11Z","lastTransitionTime":"2026-01-26T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.118681 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.118752 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.118761 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.118778 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.118791 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:11Z","lastTransitionTime":"2026-01-26T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.176211 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-09 08:33:03.755899859 +0000 UTC Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.183726 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.183862 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.183762 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.183937 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:11 crc kubenswrapper[4872]: E0126 09:09:11.183990 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:11 crc kubenswrapper[4872]: E0126 09:09:11.184169 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:11 crc kubenswrapper[4872]: E0126 09:09:11.184214 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:11 crc kubenswrapper[4872]: E0126 09:09:11.184281 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.221965 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.222271 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.222371 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.222463 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.222559 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:11Z","lastTransitionTime":"2026-01-26T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.325052 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.325565 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.325849 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.326073 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.326265 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:11Z","lastTransitionTime":"2026-01-26T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.429100 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.429163 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.429181 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.429206 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.429225 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:11Z","lastTransitionTime":"2026-01-26T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.532546 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.532629 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.532648 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.532676 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.532694 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:11Z","lastTransitionTime":"2026-01-26T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.635749 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.635828 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.635846 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.635867 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.635881 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:11Z","lastTransitionTime":"2026-01-26T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.738731 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.738845 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.738867 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.738898 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.738919 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:11Z","lastTransitionTime":"2026-01-26T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.842610 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.842671 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.842689 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.842716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.842736 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:11Z","lastTransitionTime":"2026-01-26T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.946075 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.946186 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.946206 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.946274 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:11 crc kubenswrapper[4872]: I0126 09:09:11.946295 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:11Z","lastTransitionTime":"2026-01-26T09:09:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.049733 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.049794 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.049843 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.049874 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.049890 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:12Z","lastTransitionTime":"2026-01-26T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.153193 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.153245 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.153257 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.153280 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.153294 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:12Z","lastTransitionTime":"2026-01-26T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.177001 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 14:40:31.659314582 +0000 UTC Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.184114 4872 scope.go:117] "RemoveContainer" containerID="e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621" Jan 26 09:09:12 crc kubenswrapper[4872]: E0126 09:09:12.184294 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6p5bj_openshift-ovn-kubernetes(e2577331-1886-4abb-896b-8d8a112e40d8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.201408 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2f94b985-1dcb-4c1a-acd4-fb832a143062\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef5171226ac354b2145fa9892979ebe8ac5512759c229ad2b712c41a3ddf276c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://64bf3200961007bb4670f274ea915898cae46c789706b036bbcc2447ea52f8de\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://254d28c7a95ac32b9dca328a5a0d7472431d14f4fb2b75dd3ca44944e3a62a5c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.232517 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1742750b-48f5-4949-a484-098b993387a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8666f4444ad956be5d0f9df50ab7bed18fd4329d0340878d9ed90c2f6caedae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dda455ffcb6006b4d54b3c2a3409c20d906a20383387b81d9213f7977b6687c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2066dee47fab2ecdfc4f1ae47325e990f49ad5e187d61dd117c7296dce656be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ece76e064f9f38c88c42ff1eac5339198f170d05566b45b0afe70fc8bf51568\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ebd8e3cec156dc548fe51d9c8bc31692c17c5771ef734f456b149e39b38f2e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7d7efec8f73efb5054db56663360a2ec3c4da200ce4fa667c452b4130b1063d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a6c2cd58fc7f4ec4f6bb0ef8a0746851fbec5e31b005aa783bb547d57aded64\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://82bae6ba31991b00bf31521ccb9759064f8fae977f1015716c2d9460f1857ad4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.252097 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-swcf5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3579107d-a172-4905-a491-6dd2b5a304a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4960fa5dfa153921fd4c32dffe9662db8304b2a8242ef2b570ccc174c76279d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:54Z\\\",\\\"message\\\":\\\"2026-01-26T09:08:09+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_04bb5e40-547e-4db5-9678-a80684512fdc\\\\n2026-01-26T09:08:09+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_04bb5e40-547e-4db5-9678-a80684512fdc to /host/opt/cni/bin/\\\\n2026-01-26T09:08:09Z [verbose] multus-daemon started\\\\n2026-01-26T09:08:09Z [verbose] Readiness Indicator file check\\\\n2026-01-26T09:08:54Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gbs94\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-swcf5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.256837 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.256901 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.256923 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.256953 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.256971 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:12Z","lastTransitionTime":"2026-01-26T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.296558 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-6xjb8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"32b89c0c-4fda-4490-afe4-bf441dd3d337\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8addd3b104624ae01afe58edd941442d1a596d783e2a69393477e6816747558d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pn756\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-6xjb8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.325330 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"87729ec8-51c9-4c4f-b49c-d1201b96939f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3ffe6e08bb7e39a4bd7589c5b328578c6168f7563497d52bb47aabf2867996f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://546d62a920df516634c0ba6c2cd8d78d4bbb1d08f602d9459d651e9198563722\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jv8m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:19Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-8h9pz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.340262 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a28903ba-307e-40de-9adf-9ff4cb83731d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72ce2232b80febe7ce3f28e689b9ac4b270d6c623038a1eb0e65a336d4e9aa34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bc1cf165c53ff3c104af93fdeff6c89169e4b6acf7ea669c4a30468359610483\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc1cf165c53ff3c104af93fdeff6c89169e4b6acf7ea669c4a30468359610483\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.353556 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.359134 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.359190 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.359201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.359222 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.359234 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:12Z","lastTransitionTime":"2026-01-26T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.367487 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b95e2674576bbb02bdcba8151d673ccfee7f715470fb66e1ba0c0dacd35f184f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.382286 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef70258d52e4d16ef4cd76b74e92ca2dcd6941e83613a30234d954cd0bc80d81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e1e7e1423898ece9435c0b5a24e4d3c67daaf5adf5f2f35fcf577e4f1f159b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.397596 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.410430 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fca44d96-a000-4bf2-8283-a937b0192880\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec3d9f351d9a5bdee325cf6d6178df6ebb79c954302c1d8579c80dcc73b157aa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gkwmd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-gt4gn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.423450 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krd2b" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f34a04d-444c-4640-9b51-51850a6317e6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:21Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2whh7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:21Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krd2b\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.437712 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13bb0fc1-b822-4651-ab5f-b095c86e936b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1754e95c65f24f7086f23e38595f8d3b6acb9baaaed809afee607eb5a13d06bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://178c8017e99cb8783654f3249f3499f6e61c7cabfbedf1bd078e3f345b3e7cb8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3c7b1aa5169eb070e390f0b0e5f06149b73145a6a041e0c3e6ba46857f77814\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51c595dcb014bd58ed9b48d5b4336cbe6626462b7a34e5f141d92266dde7308e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.459036 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2577331-1886-4abb-896b-8d8a112e40d8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-01-26T09:08:58Z\\\",\\\"message\\\":\\\"ager: error found while processing openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk: failed to check if pod openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk is in primary UDN: could not find OVN pod annotation in map[]\\\\nI0126 09:08:57.995458 6522 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-dns-operator/dns-operator-744455d44c-w5lkb: failed to check if pod openshift-dns-operator/dns-operator-744455d44c-w5lkb is in primary UDN: could not find OVN pod annotation in map[]\\\\nI0126 09:08:57.995521 6522 controller.go:257] Controller udn-host-isolation-manager: error found while processing openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7: failed to check if pod openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7 is in primary UDN: could not find OVN pod annotation in map[openshift.io/required-scc:restricted-v2 openshift.io/scc:restricted-v2 seccomp.security.alpha.kubernetes.io/pod:runtime/default]\\\\nE0126 09:08:58.086544 6522 shared_informer.go:316] \\\\\\\"Unhandled Error\\\\\\\" err=\\\\\\\"unable to sync caches for ovn-lb-controller\\\\\\\" logger=\\\\\\\"UnhandledError\\\\\\\"\\\\nI0126 09:08:58.087737 6522 ovnkube.go:599] Stopped ovnkube\\\\nI0126 09:08:58.087837 6522 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-6p5bj_openshift-ovn-kubernetes(e2577331-1886-4abb-896b-8d8a112e40d8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qtxcp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-6p5bj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.462030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.462119 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.462131 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.462150 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.462163 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:12Z","lastTransitionTime":"2026-01-26T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.472422 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-d66vd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5461db0d-f9c7-4215-b259-9e10e2d318df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e47b1972c41cd1d4518d37c26ecd714c580f0aaf6f275e75add04a9c745e7ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x2c7c\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-d66vd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.486424 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24e9e28-d881-4922-9462-87bccec0c844\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:07:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI0126 09:07:59.763413 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0126 09:07:59.774935 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-373237936/tls.crt::/tmp/serving-cert-373237936/tls.key\\\\\\\"\\\\nI0126 09:08:05.085530 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0126 09:08:05.099563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0126 09:08:05.099631 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0126 09:08:05.099694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0126 09:08:05.099706 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0126 09:08:05.109840 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0126 09:08:05.109871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0126 09:08:05.109883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0126 09:08:05.109888 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0126 09:08:05.109892 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0126 09:08:05.109897 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0126 09:08:05.110081 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0126 09:08:05.112015 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:07:49Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:07:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:07:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:07:47Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.499686 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d25366b4dca21fbc92318defde54dbbf003b4219a4462fda90b53a09c4b5e29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.511707 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.528029 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"732e890d-6c77-4ffa-b74b-1f129c6a96ab\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-01-26T09:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a19cd303109a35fd9522b5f6275fe0fd3425e3fdc45229d7cc65271b21f61626\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-01-26T09:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8819e3d1c1f784651df624ac6d73cfd87212c301d19c46900c64f8c4f301779e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f2fee0ff9d23a692c24c8b38e70e53b3cf5483d24041469b7f3c221b6ca3dd30\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5ba8a425ce13af2c4ff619a5e1e9bc25dc0bd8ec405df1520445145fae195c2c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://648c5b85822d0720f9153cb6d73a4ea46a32e38ba25ba6bf9f25d6d8e6e1b43a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://79182dd2b2e103505c456ca2d5c1e62157cbfc4fec894c14e41196ca7bc8c9fb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://065fce3b3c1f534ed7d8f885a8a25eb0835a079e4248238cd23d3e97cd900aac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-01-26T09:08:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-01-26T09:08:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ldz52\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-01-26T09:08:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-5sxjr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-01-26T09:09:12Z is after 2025-08-24T17:21:41Z" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.564536 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.564669 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.564737 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.564831 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.564890 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:12Z","lastTransitionTime":"2026-01-26T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.667706 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.667749 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.667762 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.667780 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.667793 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:12Z","lastTransitionTime":"2026-01-26T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.770403 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.770447 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.770458 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.770474 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.770486 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:12Z","lastTransitionTime":"2026-01-26T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.873120 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.873191 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.873217 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.873249 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.873290 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:12Z","lastTransitionTime":"2026-01-26T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.975912 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.975969 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.975982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.976002 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:12 crc kubenswrapper[4872]: I0126 09:09:12.976013 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:12Z","lastTransitionTime":"2026-01-26T09:09:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.078528 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.078572 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.078585 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.078600 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.078611 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:13Z","lastTransitionTime":"2026-01-26T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.178021 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 03:24:42.274201504 +0000 UTC Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.181768 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.182157 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.182233 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.182299 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.182422 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:13Z","lastTransitionTime":"2026-01-26T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.182892 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.182916 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.182961 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.182927 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:13 crc kubenswrapper[4872]: E0126 09:09:13.183008 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:13 crc kubenswrapper[4872]: E0126 09:09:13.183188 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:13 crc kubenswrapper[4872]: E0126 09:09:13.183282 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:13 crc kubenswrapper[4872]: E0126 09:09:13.183427 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.285090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.285143 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.285157 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.285180 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.285194 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:13Z","lastTransitionTime":"2026-01-26T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.388407 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.388491 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.388511 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.388541 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.388560 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:13Z","lastTransitionTime":"2026-01-26T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.491665 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.491714 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.491725 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.491748 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.491762 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:13Z","lastTransitionTime":"2026-01-26T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.593766 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.593836 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.593863 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.593885 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.593901 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:13Z","lastTransitionTime":"2026-01-26T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.696234 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.696299 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.696323 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.696355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.696377 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:13Z","lastTransitionTime":"2026-01-26T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.799209 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.799277 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.799295 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.799324 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.799344 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:13Z","lastTransitionTime":"2026-01-26T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.902422 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.902478 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.902490 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.902512 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:13 crc kubenswrapper[4872]: I0126 09:09:13.902526 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:13Z","lastTransitionTime":"2026-01-26T09:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.005147 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.005212 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.005223 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.005242 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.005254 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:14Z","lastTransitionTime":"2026-01-26T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.109703 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.109769 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.109786 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.109830 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.109854 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:14Z","lastTransitionTime":"2026-01-26T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.179114 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 12:20:01.903997936 +0000 UTC Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.213219 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.213315 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.213324 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.213346 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.213357 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:14Z","lastTransitionTime":"2026-01-26T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.315708 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.316400 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.316573 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.316716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.316882 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:14Z","lastTransitionTime":"2026-01-26T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.419950 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.420019 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.420042 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.420091 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.420120 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:14Z","lastTransitionTime":"2026-01-26T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.524005 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.524406 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.524564 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.524720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.524912 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:14Z","lastTransitionTime":"2026-01-26T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.627984 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.628389 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.628477 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.628554 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.628615 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:14Z","lastTransitionTime":"2026-01-26T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.732304 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.732383 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.732395 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.732420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.732435 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:14Z","lastTransitionTime":"2026-01-26T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.836350 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.836838 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.836976 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.837255 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.837347 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:14Z","lastTransitionTime":"2026-01-26T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.940506 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.940599 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.940623 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.940658 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:14 crc kubenswrapper[4872]: I0126 09:09:14.940685 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:14Z","lastTransitionTime":"2026-01-26T09:09:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.044219 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.044580 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.044696 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.044829 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.044969 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:15Z","lastTransitionTime":"2026-01-26T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.148225 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.148662 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.148741 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.148884 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.149019 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:15Z","lastTransitionTime":"2026-01-26T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.180249 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 14:16:41.083479061 +0000 UTC Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.183662 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.183911 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.183917 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:15 crc kubenswrapper[4872]: E0126 09:09:15.184170 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.184202 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:15 crc kubenswrapper[4872]: E0126 09:09:15.184436 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:15 crc kubenswrapper[4872]: E0126 09:09:15.184613 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:15 crc kubenswrapper[4872]: E0126 09:09:15.184491 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.252702 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.252784 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.252841 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.252869 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.252892 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:15Z","lastTransitionTime":"2026-01-26T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.355963 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.356025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.356049 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.356079 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.356108 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:15Z","lastTransitionTime":"2026-01-26T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.459562 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.459642 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.459667 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.459700 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.459723 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:15Z","lastTransitionTime":"2026-01-26T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.562495 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.562564 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.562589 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.562621 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.562644 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:15Z","lastTransitionTime":"2026-01-26T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.665874 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.665911 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.665923 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.665939 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.665950 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:15Z","lastTransitionTime":"2026-01-26T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.769131 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.769356 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.769365 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.769381 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.769390 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:15Z","lastTransitionTime":"2026-01-26T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.871938 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.872009 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.872028 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.872053 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.872104 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:15Z","lastTransitionTime":"2026-01-26T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.975554 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.975607 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.975623 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.975649 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:15 crc kubenswrapper[4872]: I0126 09:09:15.975667 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:15Z","lastTransitionTime":"2026-01-26T09:09:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.079212 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.079273 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.079288 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.079312 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.079327 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:16Z","lastTransitionTime":"2026-01-26T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.180457 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 14:56:52.203653714 +0000 UTC Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.183075 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.183136 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.183149 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.183173 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.183188 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:16Z","lastTransitionTime":"2026-01-26T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.286773 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.286837 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.286849 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.286869 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.286880 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:16Z","lastTransitionTime":"2026-01-26T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.389926 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.389967 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.389977 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.389991 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.390002 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:16Z","lastTransitionTime":"2026-01-26T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.492623 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.492665 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.492676 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.492691 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.492705 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:16Z","lastTransitionTime":"2026-01-26T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.596732 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.596926 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.596956 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.596984 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.597003 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:16Z","lastTransitionTime":"2026-01-26T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.699926 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.699993 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.700005 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.700028 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.700041 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:16Z","lastTransitionTime":"2026-01-26T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.802343 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.802476 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.802493 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.802519 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.802541 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:16Z","lastTransitionTime":"2026-01-26T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.906380 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.906427 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.906436 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.906455 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:16 crc kubenswrapper[4872]: I0126 09:09:16.906466 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:16Z","lastTransitionTime":"2026-01-26T09:09:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.009376 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.009421 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.009434 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.009469 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.009482 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:17Z","lastTransitionTime":"2026-01-26T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.113474 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.113515 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.113524 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.113536 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.113546 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:17Z","lastTransitionTime":"2026-01-26T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.181407 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 01:19:39.154538366 +0000 UTC Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.183881 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:17 crc kubenswrapper[4872]: E0126 09:09:17.184034 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.184085 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.184138 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:17 crc kubenswrapper[4872]: E0126 09:09:17.184160 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:17 crc kubenswrapper[4872]: E0126 09:09:17.184346 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.184407 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:17 crc kubenswrapper[4872]: E0126 09:09:17.184535 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.215737 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.215856 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.215877 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.215902 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.215922 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:17Z","lastTransitionTime":"2026-01-26T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.276880 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podStartSLOduration=71.276856615 podStartE2EDuration="1m11.276856615s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:17.260341962 +0000 UTC m=+90.569181763" watchObservedRunningTime="2026-01-26 09:09:17.276856615 +0000 UTC m=+90.585696426" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.297700 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-8h9pz" podStartSLOduration=71.297680738 podStartE2EDuration="1m11.297680738s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:17.297224344 +0000 UTC m=+90.606064145" watchObservedRunningTime="2026-01-26 09:09:17.297680738 +0000 UTC m=+90.606520539" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.298165 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-6xjb8" podStartSLOduration=71.2981569 podStartE2EDuration="1m11.2981569s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:17.279411116 +0000 UTC m=+90.588250907" watchObservedRunningTime="2026-01-26 09:09:17.2981569 +0000 UTC m=+90.606996701" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.310605 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=21.310583132 podStartE2EDuration="21.310583132s" podCreationTimestamp="2026-01-26 09:08:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:17.310232062 +0000 UTC m=+90.619071863" watchObservedRunningTime="2026-01-26 09:09:17.310583132 +0000 UTC m=+90.619422933" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.319005 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.319239 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.319368 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.319672 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.319978 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:17Z","lastTransitionTime":"2026-01-26T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.359362 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-d66vd" podStartSLOduration=71.359342791 podStartE2EDuration="1m11.359342791s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:17.345007947 +0000 UTC m=+90.653847748" watchObservedRunningTime="2026-01-26 09:09:17.359342791 +0000 UTC m=+90.668182592" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.402967 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=40.402939958 podStartE2EDuration="40.402939958s" podCreationTimestamp="2026-01-26 09:08:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:17.359821924 +0000 UTC m=+90.668661725" watchObservedRunningTime="2026-01-26 09:09:17.402939958 +0000 UTC m=+90.711779769" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.422286 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.422326 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.422338 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.422355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.422366 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:17Z","lastTransitionTime":"2026-01-26T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.442238 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-5sxjr" podStartSLOduration=71.442219697 podStartE2EDuration="1m11.442219697s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:17.441301072 +0000 UTC m=+90.750140873" watchObservedRunningTime="2026-01-26 09:09:17.442219697 +0000 UTC m=+90.751059498" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.487119 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=72.487088729 podStartE2EDuration="1m12.487088729s" podCreationTimestamp="2026-01-26 09:08:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:17.463417039 +0000 UTC m=+90.772256860" watchObservedRunningTime="2026-01-26 09:09:17.487088729 +0000 UTC m=+90.795928570" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.505982 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-swcf5" podStartSLOduration=71.505961287 podStartE2EDuration="1m11.505961287s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:17.505433163 +0000 UTC m=+90.814272974" watchObservedRunningTime="2026-01-26 09:09:17.505961287 +0000 UTC m=+90.814801098" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.525131 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.525384 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.525498 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.525579 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.525748 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:17Z","lastTransitionTime":"2026-01-26T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.528126 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=70.528113626 podStartE2EDuration="1m10.528113626s" podCreationTimestamp="2026-01-26 09:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:17.527862879 +0000 UTC m=+90.836702680" watchObservedRunningTime="2026-01-26 09:09:17.528113626 +0000 UTC m=+90.836953427" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.556219 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=70.556190387 podStartE2EDuration="1m10.556190387s" podCreationTimestamp="2026-01-26 09:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:17.554129251 +0000 UTC m=+90.862969052" watchObservedRunningTime="2026-01-26 09:09:17.556190387 +0000 UTC m=+90.865030208" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.628365 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.628661 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.628743 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.628863 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.628944 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:17Z","lastTransitionTime":"2026-01-26T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.732103 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.732175 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.732193 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.732220 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.732244 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:17Z","lastTransitionTime":"2026-01-26T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.836185 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.836226 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.836237 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.836252 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.836262 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:17Z","lastTransitionTime":"2026-01-26T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.939275 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.939680 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.939841 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.939993 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:17 crc kubenswrapper[4872]: I0126 09:09:17.940093 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:17Z","lastTransitionTime":"2026-01-26T09:09:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.044687 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.045270 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.045501 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.045694 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.045888 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:18Z","lastTransitionTime":"2026-01-26T09:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.149727 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.149781 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.149792 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.149833 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.149844 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:18Z","lastTransitionTime":"2026-01-26T09:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.182630 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-26 15:04:36.582055164 +0000 UTC Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.252446 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.252497 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.252509 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.252525 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.252536 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:18Z","lastTransitionTime":"2026-01-26T09:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.356061 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.356129 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.356141 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.356162 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.356174 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:18Z","lastTransitionTime":"2026-01-26T09:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.459753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.459856 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.459885 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.459916 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.459940 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:18Z","lastTransitionTime":"2026-01-26T09:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.563424 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.563483 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.563494 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.563509 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.563521 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:18Z","lastTransitionTime":"2026-01-26T09:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.668776 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.668868 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.668890 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.668918 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.668937 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:18Z","lastTransitionTime":"2026-01-26T09:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.772125 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.772186 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.772199 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.772226 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.772242 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:18Z","lastTransitionTime":"2026-01-26T09:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.876090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.876156 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.876178 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.876208 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.876230 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:18Z","lastTransitionTime":"2026-01-26T09:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.978631 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.978678 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.978696 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.978717 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:18 crc kubenswrapper[4872]: I0126 09:09:18.978733 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:18Z","lastTransitionTime":"2026-01-26T09:09:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.081416 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.081475 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.081487 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.081506 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.081520 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:19Z","lastTransitionTime":"2026-01-26T09:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.182969 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:19 crc kubenswrapper[4872]: E0126 09:09:19.183100 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.183215 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.183222 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.183378 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:19 crc kubenswrapper[4872]: E0126 09:09:19.183565 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:19 crc kubenswrapper[4872]: E0126 09:09:19.183949 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.184011 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 22:54:22.300432843 +0000 UTC Jan 26 09:09:19 crc kubenswrapper[4872]: E0126 09:09:19.184134 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.185586 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.185652 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.185669 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.185693 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.185710 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:19Z","lastTransitionTime":"2026-01-26T09:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.287958 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.288010 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.288028 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.288054 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.288071 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:19Z","lastTransitionTime":"2026-01-26T09:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.391250 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.391312 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.391324 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.391342 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.391358 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:19Z","lastTransitionTime":"2026-01-26T09:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.493698 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.493758 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.493770 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.493788 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.493819 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:19Z","lastTransitionTime":"2026-01-26T09:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.596966 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.597044 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.597057 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.597080 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.597094 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:19Z","lastTransitionTime":"2026-01-26T09:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.699910 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.699942 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.699952 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.699968 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.699984 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:19Z","lastTransitionTime":"2026-01-26T09:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.802680 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.803090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.803232 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.803435 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.803683 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:19Z","lastTransitionTime":"2026-01-26T09:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.911089 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.911529 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.911668 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.911796 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:19 crc kubenswrapper[4872]: I0126 09:09:19.912005 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:19Z","lastTransitionTime":"2026-01-26T09:09:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.015086 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.015190 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.015217 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.015247 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.015269 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:20Z","lastTransitionTime":"2026-01-26T09:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.117884 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.117958 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.117981 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.118014 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.118037 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:20Z","lastTransitionTime":"2026-01-26T09:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.184535 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-01 02:59:22.548094876 +0000 UTC Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.220545 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.220578 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.220588 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.220604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.220616 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:20Z","lastTransitionTime":"2026-01-26T09:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.249282 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.249344 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.249354 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.249386 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.249399 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-01-26T09:09:20Z","lastTransitionTime":"2026-01-26T09:09:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.311098 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl"] Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.311654 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.314151 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.314235 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.314615 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.314742 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.406197 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ebce20f-174c-420e-9dac-c82a3133d020-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.406280 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7ebce20f-174c-420e-9dac-c82a3133d020-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.406359 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7ebce20f-174c-420e-9dac-c82a3133d020-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.406390 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ebce20f-174c-420e-9dac-c82a3133d020-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.406445 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7ebce20f-174c-420e-9dac-c82a3133d020-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.507725 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ebce20f-174c-420e-9dac-c82a3133d020-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.508306 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7ebce20f-174c-420e-9dac-c82a3133d020-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.508570 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7ebce20f-174c-420e-9dac-c82a3133d020-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.508872 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ebce20f-174c-420e-9dac-c82a3133d020-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.509131 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7ebce20f-174c-420e-9dac-c82a3133d020-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.508385 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/7ebce20f-174c-420e-9dac-c82a3133d020-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.508674 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/7ebce20f-174c-420e-9dac-c82a3133d020-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.510789 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7ebce20f-174c-420e-9dac-c82a3133d020-service-ca\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.522599 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7ebce20f-174c-420e-9dac-c82a3133d020-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.540637 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ebce20f-174c-420e-9dac-c82a3133d020-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-9bpdl\" (UID: \"7ebce20f-174c-420e-9dac-c82a3133d020\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.633825 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" Jan 26 09:09:20 crc kubenswrapper[4872]: I0126 09:09:20.799184 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" event={"ID":"7ebce20f-174c-420e-9dac-c82a3133d020","Type":"ContainerStarted","Data":"918887b7d98ff6059521620b86e5034056f9cba093edad0a246d188905cef366"} Jan 26 09:09:21 crc kubenswrapper[4872]: I0126 09:09:21.183106 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:21 crc kubenswrapper[4872]: I0126 09:09:21.183145 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:21 crc kubenswrapper[4872]: E0126 09:09:21.183703 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:21 crc kubenswrapper[4872]: E0126 09:09:21.183778 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:21 crc kubenswrapper[4872]: I0126 09:09:21.183198 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:21 crc kubenswrapper[4872]: I0126 09:09:21.183269 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:21 crc kubenswrapper[4872]: E0126 09:09:21.184118 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:21 crc kubenswrapper[4872]: E0126 09:09:21.183943 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:21 crc kubenswrapper[4872]: I0126 09:09:21.186579 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 23:30:20.641392997 +0000 UTC Jan 26 09:09:21 crc kubenswrapper[4872]: I0126 09:09:21.186658 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Jan 26 09:09:21 crc kubenswrapper[4872]: I0126 09:09:21.196334 4872 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Jan 26 09:09:21 crc kubenswrapper[4872]: I0126 09:09:21.805627 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" event={"ID":"7ebce20f-174c-420e-9dac-c82a3133d020","Type":"ContainerStarted","Data":"d476b3aab37055ab8b93f94a20156f96c206e7f4c37c920b0e30fab71ade195b"} Jan 26 09:09:23 crc kubenswrapper[4872]: I0126 09:09:23.183475 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:23 crc kubenswrapper[4872]: I0126 09:09:23.183468 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:23 crc kubenswrapper[4872]: I0126 09:09:23.183578 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:23 crc kubenswrapper[4872]: E0126 09:09:23.183759 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:23 crc kubenswrapper[4872]: I0126 09:09:23.183811 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:23 crc kubenswrapper[4872]: E0126 09:09:23.184116 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:23 crc kubenswrapper[4872]: E0126 09:09:23.184595 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:23 crc kubenswrapper[4872]: E0126 09:09:23.184734 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:23 crc kubenswrapper[4872]: I0126 09:09:23.185136 4872 scope.go:117] "RemoveContainer" containerID="e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621" Jan 26 09:09:23 crc kubenswrapper[4872]: I0126 09:09:23.816161 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovnkube-controller/2.log" Jan 26 09:09:23 crc kubenswrapper[4872]: I0126 09:09:23.819655 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerStarted","Data":"41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed"} Jan 26 09:09:23 crc kubenswrapper[4872]: I0126 09:09:23.821057 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:09:23 crc kubenswrapper[4872]: I0126 09:09:23.859582 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-9bpdl" podStartSLOduration=77.85956248 podStartE2EDuration="1m17.85956248s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:21.825267158 +0000 UTC m=+95.134106959" watchObservedRunningTime="2026-01-26 09:09:23.85956248 +0000 UTC m=+97.168402291" Jan 26 09:09:23 crc kubenswrapper[4872]: I0126 09:09:23.859877 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podStartSLOduration=77.859869099 podStartE2EDuration="1m17.859869099s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:23.859504389 +0000 UTC m=+97.168344200" watchObservedRunningTime="2026-01-26 09:09:23.859869099 +0000 UTC m=+97.168708910" Jan 26 09:09:24 crc kubenswrapper[4872]: I0126 09:09:24.209004 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-krd2b"] Jan 26 09:09:24 crc kubenswrapper[4872]: I0126 09:09:24.209158 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:24 crc kubenswrapper[4872]: E0126 09:09:24.209302 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:25 crc kubenswrapper[4872]: I0126 09:09:25.079500 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:25 crc kubenswrapper[4872]: E0126 09:09:25.079759 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:09:25 crc kubenswrapper[4872]: E0126 09:09:25.080190 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs podName:8f34a04d-444c-4640-9b51-51850a6317e6 nodeName:}" failed. No retries permitted until 2026-01-26 09:10:29.080163956 +0000 UTC m=+162.389003797 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs") pod "network-metrics-daemon-krd2b" (UID: "8f34a04d-444c-4640-9b51-51850a6317e6") : object "openshift-multus"/"metrics-daemon-secret" not registered Jan 26 09:09:25 crc kubenswrapper[4872]: I0126 09:09:25.183403 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:25 crc kubenswrapper[4872]: I0126 09:09:25.183503 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:25 crc kubenswrapper[4872]: E0126 09:09:25.183622 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Jan 26 09:09:25 crc kubenswrapper[4872]: I0126 09:09:25.183732 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:25 crc kubenswrapper[4872]: E0126 09:09:25.184106 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Jan 26 09:09:25 crc kubenswrapper[4872]: E0126 09:09:25.184298 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.183461 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:26 crc kubenswrapper[4872]: E0126 09:09:26.183672 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krd2b" podUID="8f34a04d-444c-4640-9b51-51850a6317e6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.442081 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.442337 4872 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.507354 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kxnj5"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.507974 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.508690 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zw7z6"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.509300 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.510453 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-x9zlw"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.511020 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.511502 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.512116 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.514914 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.515127 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.515487 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.515663 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.515878 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-82cml"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.515926 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.516292 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.516615 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.516755 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.519626 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.519665 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.519738 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.519960 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.520079 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.520254 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.520506 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.520550 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.520666 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.520759 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.520973 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.521101 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.521192 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.521333 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.521551 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.521572 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.531974 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.539542 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.542619 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.542952 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.568707 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.582596 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.582903 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.583108 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.583461 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.583618 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.586621 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.588894 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.589420 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.589937 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.590281 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.590854 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.591343 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.592702 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.593252 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.596301 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.597646 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mznfc"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.598565 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.601916 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.606425 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.607179 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.607424 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.607526 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mk4bd"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.607734 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.607866 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.607939 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.608321 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.608888 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.608981 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.610632 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-node-pullsecrets\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.610683 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/07ffe425-ff10-4c9f-b3c5-fc2c5e167f67-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-x9zlw\" (UID: \"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.610713 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-audit-dir\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.610736 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cde35d3c-8bd5-4e20-98de-2eca40d2a38c-config\") pod \"console-operator-58897d9998-zw7z6\" (UID: \"cde35d3c-8bd5-4e20-98de-2eca40d2a38c\") " pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.610761 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zscxs\" (UniqueName: \"kubernetes.io/projected/cde35d3c-8bd5-4e20-98de-2eca40d2a38c-kube-api-access-zscxs\") pod \"console-operator-58897d9998-zw7z6\" (UID: \"cde35d3c-8bd5-4e20-98de-2eca40d2a38c\") " pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.610784 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-audit\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.610841 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8486391-70e4-4b2f-a093-ea9ab42c18bd-serving-cert\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.610865 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3df101-99e3-41a0-a235-0876a7249434-serving-cert\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.610929 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ffe425-ff10-4c9f-b3c5-fc2c5e167f67-config\") pod \"machine-api-operator-5694c8668f-x9zlw\" (UID: \"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.610961 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f8486391-70e4-4b2f-a093-ea9ab42c18bd-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611033 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cde35d3c-8bd5-4e20-98de-2eca40d2a38c-trusted-ca\") pod \"console-operator-58897d9998-zw7z6\" (UID: \"cde35d3c-8bd5-4e20-98de-2eca40d2a38c\") " pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611076 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611101 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdspk\" (UniqueName: \"kubernetes.io/projected/f8486391-70e4-4b2f-a093-ea9ab42c18bd-kube-api-access-zdspk\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611127 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f8486391-70e4-4b2f-a093-ea9ab42c18bd-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611150 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cde35d3c-8bd5-4e20-98de-2eca40d2a38c-serving-cert\") pod \"console-operator-58897d9998-zw7z6\" (UID: \"cde35d3c-8bd5-4e20-98de-2eca40d2a38c\") " pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611175 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-etcd-client\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611205 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-image-import-ca\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611235 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-serving-cert\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611261 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f8486391-70e4-4b2f-a093-ea9ab42c18bd-audit-policies\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611297 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz9xr\" (UniqueName: \"kubernetes.io/projected/3a3df101-99e3-41a0-a235-0876a7249434-kube-api-access-zz9xr\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611324 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/07ffe425-ff10-4c9f-b3c5-fc2c5e167f67-images\") pod \"machine-api-operator-5694c8668f-x9zlw\" (UID: \"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611358 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-etcd-serving-ca\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611404 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-config\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611433 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611475 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f8486391-70e4-4b2f-a093-ea9ab42c18bd-audit-dir\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611504 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msbxx\" (UniqueName: \"kubernetes.io/projected/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-kube-api-access-msbxx\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611533 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f8486391-70e4-4b2f-a093-ea9ab42c18bd-encryption-config\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611568 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-encryption-config\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611601 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-client-ca\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611631 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-config\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611671 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrxkv\" (UniqueName: \"kubernetes.io/projected/07ffe425-ff10-4c9f-b3c5-fc2c5e167f67-kube-api-access-hrxkv\") pod \"machine-api-operator-5694c8668f-x9zlw\" (UID: \"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.611699 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f8486391-70e4-4b2f-a093-ea9ab42c18bd-etcd-client\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.618353 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.618394 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hx6pl"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.618640 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.618815 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.618968 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.619103 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.619198 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.619227 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.619343 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.619878 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.620003 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.620322 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.620910 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p975q"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.621934 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.626732 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.627600 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.629009 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.629717 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.631310 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.632111 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.632760 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-w5lkb"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.633734 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-w5lkb" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.635197 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-zr6z9"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.641584 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-ztd4t"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.641854 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-zr6z9" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.641991 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.642256 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.642822 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.643224 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.643718 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.664726 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.665151 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.678417 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.679398 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.679634 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.697998 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.698857 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.701168 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.702014 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-n7pfp"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.702858 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.703522 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.703841 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.703859 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.704030 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.704860 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.704921 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.705037 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.705240 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.705467 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.705521 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.705644 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.705651 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.705791 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.705832 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.705944 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.706082 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.706103 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.706326 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.706538 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.706654 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.706745 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.710962 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.711303 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.711380 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.711550 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.712771 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-serving-cert\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.712842 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-config\") pod \"route-controller-manager-6576b87f9c-pzn5w\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.712867 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5dgh7\" (UID: \"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.712894 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d32d9b8b-2373-4c15-b5ce-f7f77c899cdd-srv-cert\") pod \"catalog-operator-68c6474976-6bfrv\" (UID: \"d32d9b8b-2373-4c15-b5ce-f7f77c899cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.712915 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ebeadae0-aa3e-47d1-8068-1c339da55d91-proxy-tls\") pod \"machine-config-operator-74547568cd-6gqpg\" (UID: \"ebeadae0-aa3e-47d1-8068-1c339da55d91\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.712941 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a111e46-9f4e-465f-821f-b58864174417-serving-cert\") pod \"openshift-config-operator-7777fb866f-mznfc\" (UID: \"5a111e46-9f4e-465f-821f-b58864174417\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.713005 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.713054 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdspk\" (UniqueName: \"kubernetes.io/projected/f8486391-70e4-4b2f-a093-ea9ab42c18bd-kube-api-access-zdspk\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.713116 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cde35d3c-8bd5-4e20-98de-2eca40d2a38c-trusted-ca\") pod \"console-operator-58897d9998-zw7z6\" (UID: \"cde35d3c-8bd5-4e20-98de-2eca40d2a38c\") " pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.713170 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/52b51ab7-5c1b-4cbb-a471-905a56f96a84-metrics-tls\") pod \"dns-operator-744455d44c-w5lkb\" (UID: \"52b51ab7-5c1b-4cbb-a471-905a56f96a84\") " pod="openshift-dns-operator/dns-operator-744455d44c-w5lkb" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.713221 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-audit-policies\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.713258 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.713290 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f8486391-70e4-4b2f-a093-ea9ab42c18bd-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.713315 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cde35d3c-8bd5-4e20-98de-2eca40d2a38c-serving-cert\") pod \"console-operator-58897d9998-zw7z6\" (UID: \"cde35d3c-8bd5-4e20-98de-2eca40d2a38c\") " pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.713338 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-serving-cert\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714023 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-oauth-serving-cert\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714070 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714091 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714116 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-etcd-client\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714135 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-image-import-ca\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714154 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kphgq\" (UniqueName: \"kubernetes.io/projected/586786e8-df0e-4424-8c84-bc2408b2dea5-kube-api-access-kphgq\") pod \"machine-approver-56656f9798-c4vmf\" (UID: \"586786e8-df0e-4424-8c84-bc2408b2dea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714171 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-client-ca\") pod \"route-controller-manager-6576b87f9c-pzn5w\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714189 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5a111e46-9f4e-465f-821f-b58864174417-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mznfc\" (UID: \"5a111e46-9f4e-465f-821f-b58864174417\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714214 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714251 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d32d9b8b-2373-4c15-b5ce-f7f77c899cdd-profile-collector-cert\") pod \"catalog-operator-68c6474976-6bfrv\" (UID: \"d32d9b8b-2373-4c15-b5ce-f7f77c899cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714278 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-serving-cert\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714301 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f8486391-70e4-4b2f-a093-ea9ab42c18bd-audit-policies\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714323 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-serving-cert\") pod \"route-controller-manager-6576b87f9c-pzn5w\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714353 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz9xr\" (UniqueName: \"kubernetes.io/projected/3a3df101-99e3-41a0-a235-0876a7249434-kube-api-access-zz9xr\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714392 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/07ffe425-ff10-4c9f-b3c5-fc2c5e167f67-images\") pod \"machine-api-operator-5694c8668f-x9zlw\" (UID: \"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714417 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc98b\" (UniqueName: \"kubernetes.io/projected/d32d9b8b-2373-4c15-b5ce-f7f77c899cdd-kube-api-access-dc98b\") pod \"catalog-operator-68c6474976-6bfrv\" (UID: \"d32d9b8b-2373-4c15-b5ce-f7f77c899cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714446 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-service-ca\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714468 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/596ad6ba-b5ec-471c-96b6-1b24d525b163-audit-dir\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714493 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-etcd-serving-ca\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714517 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/586786e8-df0e-4424-8c84-bc2408b2dea5-machine-approver-tls\") pod \"machine-approver-56656f9798-c4vmf\" (UID: \"586786e8-df0e-4424-8c84-bc2408b2dea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714529 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714665 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714754 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.715048 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.715174 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.705950 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.714536 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-config\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.715387 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz6lv\" (UniqueName: \"kubernetes.io/projected/f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc-kube-api-access-lz6lv\") pod \"cluster-image-registry-operator-dc59b4c8b-5dgh7\" (UID: \"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.715425 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/586786e8-df0e-4424-8c84-bc2408b2dea5-config\") pod \"machine-approver-56656f9798-c4vmf\" (UID: \"586786e8-df0e-4424-8c84-bc2408b2dea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.715445 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnj96\" (UniqueName: \"kubernetes.io/projected/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-kube-api-access-gnj96\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.715475 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ebeadae0-aa3e-47d1-8068-1c339da55d91-images\") pod \"machine-config-operator-74547568cd-6gqpg\" (UID: \"ebeadae0-aa3e-47d1-8068-1c339da55d91\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.715506 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-config\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.715529 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.715551 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89ad9943-b0ac-4b88-aaf0-bdb332b22488-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-g42sk\" (UID: \"89ad9943-b0ac-4b88-aaf0-bdb332b22488\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.715575 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrlrj\" (UniqueName: \"kubernetes.io/projected/5a111e46-9f4e-465f-821f-b58864174417-kube-api-access-rrlrj\") pod \"openshift-config-operator-7777fb866f-mznfc\" (UID: \"5a111e46-9f4e-465f-821f-b58864174417\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.716018 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f8486391-70e4-4b2f-a093-ea9ab42c18bd-audit-dir\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.716059 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8tzq\" (UniqueName: \"kubernetes.io/projected/596ad6ba-b5ec-471c-96b6-1b24d525b163-kube-api-access-l8tzq\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.716115 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqtpl\" (UniqueName: \"kubernetes.io/projected/64013929-bfea-444c-8bf9-afbc7625a690-kube-api-access-kqtpl\") pod \"openshift-controller-manager-operator-756b6f6bc6-th5xh\" (UID: \"64013929-bfea-444c-8bf9-afbc7625a690\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.716287 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msbxx\" (UniqueName: \"kubernetes.io/projected/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-kube-api-access-msbxx\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.716524 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f8486391-70e4-4b2f-a093-ea9ab42c18bd-encryption-config\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.716591 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-oauth-config\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.716662 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.716694 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/586786e8-df0e-4424-8c84-bc2408b2dea5-auth-proxy-config\") pod \"machine-approver-56656f9798-c4vmf\" (UID: \"586786e8-df0e-4424-8c84-bc2408b2dea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.716744 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-config\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.716767 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1-config\") pod \"kube-apiserver-operator-766d6c64bb-r2xzk\" (UID: \"bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.717176 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ebeadae0-aa3e-47d1-8068-1c339da55d91-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6gqpg\" (UID: \"ebeadae0-aa3e-47d1-8068-1c339da55d91\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.717224 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-encryption-config\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.717248 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-etcd-client\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.717269 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-r2xzk\" (UID: \"bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.717332 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2cpv\" (UniqueName: \"kubernetes.io/projected/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-kube-api-access-g2cpv\") pod \"route-controller-manager-6576b87f9c-pzn5w\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.717356 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.717380 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-client-ca\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.717402 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-config\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.717419 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64013929-bfea-444c-8bf9-afbc7625a690-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-th5xh\" (UID: \"64013929-bfea-444c-8bf9-afbc7625a690\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.717434 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.718148 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.718258 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.718428 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.718443 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.718560 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.718750 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-9tdst"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719195 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.717439 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89ad9943-b0ac-4b88-aaf0-bdb332b22488-config\") pod \"kube-controller-manager-operator-78b949d7b-g42sk\" (UID: \"89ad9943-b0ac-4b88-aaf0-bdb332b22488\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719554 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5dgh7\" (UID: \"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719603 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719631 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrxkv\" (UniqueName: \"kubernetes.io/projected/07ffe425-ff10-4c9f-b3c5-fc2c5e167f67-kube-api-access-hrxkv\") pod \"machine-api-operator-5694c8668f-x9zlw\" (UID: \"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719638 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8k2xh"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719655 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f8486391-70e4-4b2f-a093-ea9ab42c18bd-etcd-client\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719676 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-etcd-ca\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719703 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/89ad9943-b0ac-4b88-aaf0-bdb332b22488-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-g42sk\" (UID: \"89ad9943-b0ac-4b88-aaf0-bdb332b22488\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719726 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/07ffe425-ff10-4c9f-b3c5-fc2c5e167f67-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-x9zlw\" (UID: \"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719752 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-node-pullsecrets\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719770 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-r2xzk\" (UID: \"bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719787 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5dgh7\" (UID: \"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719827 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719849 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719903 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-audit-dir\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719974 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.720024 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-config\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.720156 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.720216 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.720319 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.720432 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.720643 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8k2xh" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.719859 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-audit-dir\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.720915 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cde35d3c-8bd5-4e20-98de-2eca40d2a38c-config\") pod \"console-operator-58897d9998-zw7z6\" (UID: \"cde35d3c-8bd5-4e20-98de-2eca40d2a38c\") " pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.720953 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-etcd-service-ca\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.720987 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rchtx\" (UniqueName: \"kubernetes.io/projected/9d625f91-d4de-4687-bdbe-8ff985d7e406-kube-api-access-rchtx\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721018 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-audit\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721043 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8486391-70e4-4b2f-a093-ea9ab42c18bd-serving-cert\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721070 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zscxs\" (UniqueName: \"kubernetes.io/projected/cde35d3c-8bd5-4e20-98de-2eca40d2a38c-kube-api-access-zscxs\") pod \"console-operator-58897d9998-zw7z6\" (UID: \"cde35d3c-8bd5-4e20-98de-2eca40d2a38c\") " pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721093 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l62xf\" (UniqueName: \"kubernetes.io/projected/52b51ab7-5c1b-4cbb-a471-905a56f96a84-kube-api-access-l62xf\") pod \"dns-operator-744455d44c-w5lkb\" (UID: \"52b51ab7-5c1b-4cbb-a471-905a56f96a84\") " pod="openshift-dns-operator/dns-operator-744455d44c-w5lkb" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721095 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721125 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3df101-99e3-41a0-a235-0876a7249434-serving-cert\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721160 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64013929-bfea-444c-8bf9-afbc7625a690-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-th5xh\" (UID: \"64013929-bfea-444c-8bf9-afbc7625a690\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721172 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f8486391-70e4-4b2f-a093-ea9ab42c18bd-audit-dir\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721190 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721226 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ffe425-ff10-4c9f-b3c5-fc2c5e167f67-config\") pod \"machine-api-operator-5694c8668f-x9zlw\" (UID: \"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721252 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-trusted-ca-bundle\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721283 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9x2w\" (UniqueName: \"kubernetes.io/projected/ebeadae0-aa3e-47d1-8068-1c339da55d91-kube-api-access-q9x2w\") pod \"machine-config-operator-74547568cd-6gqpg\" (UID: \"ebeadae0-aa3e-47d1-8068-1c339da55d91\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721321 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721353 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.721383 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f8486391-70e4-4b2f-a093-ea9ab42c18bd-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.722777 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cde35d3c-8bd5-4e20-98de-2eca40d2a38c-trusted-ca\") pod \"console-operator-58897d9998-zw7z6\" (UID: \"cde35d3c-8bd5-4e20-98de-2eca40d2a38c\") " pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.722877 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-etcd-client\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.723056 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.723600 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.723818 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-image-import-ca\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.724370 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f8486391-70e4-4b2f-a093-ea9ab42c18bd-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.725022 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f8486391-70e4-4b2f-a093-ea9ab42c18bd-encryption-config\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.725064 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.725883 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f8486391-70e4-4b2f-a093-ea9ab42c18bd-etcd-client\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.726619 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.726672 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f8486391-70e4-4b2f-a093-ea9ab42c18bd-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.726899 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-node-pullsecrets\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.727418 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-serving-cert\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.727546 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-etcd-serving-ca\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.728034 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-audit\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.728478 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/07ffe425-ff10-4c9f-b3c5-fc2c5e167f67-images\") pod \"machine-api-operator-5694c8668f-x9zlw\" (UID: \"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.729155 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07ffe425-ff10-4c9f-b3c5-fc2c5e167f67-config\") pod \"machine-api-operator-5694c8668f-x9zlw\" (UID: \"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.739004 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.739021 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-config\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.739451 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.739492 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cde35d3c-8bd5-4e20-98de-2eca40d2a38c-config\") pod \"console-operator-58897d9998-zw7z6\" (UID: \"cde35d3c-8bd5-4e20-98de-2eca40d2a38c\") " pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.740969 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.741645 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.742141 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.742461 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f8486391-70e4-4b2f-a093-ea9ab42c18bd-audit-policies\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.742562 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.742678 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.742883 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.742749 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pplqc"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.743033 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-client-ca\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.743989 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.744625 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.744790 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pplqc" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.746365 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3df101-99e3-41a0-a235-0876a7249434-serving-cert\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.746384 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cde35d3c-8bd5-4e20-98de-2eca40d2a38c-serving-cert\") pod \"console-operator-58897d9998-zw7z6\" (UID: \"cde35d3c-8bd5-4e20-98de-2eca40d2a38c\") " pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.747197 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8486391-70e4-4b2f-a093-ea9ab42c18bd-serving-cert\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.748920 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.749544 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.749694 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-encryption-config\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.752335 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.755107 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.766257 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.772175 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/07ffe425-ff10-4c9f-b3c5-fc2c5e167f67-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-x9zlw\" (UID: \"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.777294 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.780999 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.783052 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.785185 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.785967 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.795594 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.796417 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.796748 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.797746 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.800387 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bsfdw"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.801178 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.801572 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kxnj5"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.803558 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-x9zlw"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.804570 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.805317 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.806094 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ntnp2"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.806531 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.807010 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.810145 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zw7z6"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.810814 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.811903 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.815496 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-82cml"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.816777 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823011 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc98b\" (UniqueName: \"kubernetes.io/projected/d32d9b8b-2373-4c15-b5ce-f7f77c899cdd-kube-api-access-dc98b\") pod \"catalog-operator-68c6474976-6bfrv\" (UID: \"d32d9b8b-2373-4c15-b5ce-f7f77c899cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823053 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/596ad6ba-b5ec-471c-96b6-1b24d525b163-audit-dir\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823080 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-service-ca\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823108 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz6lv\" (UniqueName: \"kubernetes.io/projected/f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc-kube-api-access-lz6lv\") pod \"cluster-image-registry-operator-dc59b4c8b-5dgh7\" (UID: \"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823133 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/586786e8-df0e-4424-8c84-bc2408b2dea5-machine-approver-tls\") pod \"machine-approver-56656f9798-c4vmf\" (UID: \"586786e8-df0e-4424-8c84-bc2408b2dea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823151 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-config\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823169 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/586786e8-df0e-4424-8c84-bc2408b2dea5-config\") pod \"machine-approver-56656f9798-c4vmf\" (UID: \"586786e8-df0e-4424-8c84-bc2408b2dea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823186 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnj96\" (UniqueName: \"kubernetes.io/projected/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-kube-api-access-gnj96\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823207 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ebeadae0-aa3e-47d1-8068-1c339da55d91-images\") pod \"machine-config-operator-74547568cd-6gqpg\" (UID: \"ebeadae0-aa3e-47d1-8068-1c339da55d91\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823229 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89ad9943-b0ac-4b88-aaf0-bdb332b22488-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-g42sk\" (UID: \"89ad9943-b0ac-4b88-aaf0-bdb332b22488\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823248 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrlrj\" (UniqueName: \"kubernetes.io/projected/5a111e46-9f4e-465f-821f-b58864174417-kube-api-access-rrlrj\") pod \"openshift-config-operator-7777fb866f-mznfc\" (UID: \"5a111e46-9f4e-465f-821f-b58864174417\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823306 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8tzq\" (UniqueName: \"kubernetes.io/projected/596ad6ba-b5ec-471c-96b6-1b24d525b163-kube-api-access-l8tzq\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823325 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqtpl\" (UniqueName: \"kubernetes.io/projected/64013929-bfea-444c-8bf9-afbc7625a690-kube-api-access-kqtpl\") pod \"openshift-controller-manager-operator-756b6f6bc6-th5xh\" (UID: \"64013929-bfea-444c-8bf9-afbc7625a690\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823357 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-oauth-config\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823377 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823397 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ebeadae0-aa3e-47d1-8068-1c339da55d91-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6gqpg\" (UID: \"ebeadae0-aa3e-47d1-8068-1c339da55d91\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823438 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/586786e8-df0e-4424-8c84-bc2408b2dea5-auth-proxy-config\") pod \"machine-approver-56656f9798-c4vmf\" (UID: \"586786e8-df0e-4424-8c84-bc2408b2dea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823460 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-config\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823482 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1-config\") pod \"kube-apiserver-operator-766d6c64bb-r2xzk\" (UID: \"bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823510 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-etcd-client\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823531 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-r2xzk\" (UID: \"bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823743 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2cpv\" (UniqueName: \"kubernetes.io/projected/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-kube-api-access-g2cpv\") pod \"route-controller-manager-6576b87f9c-pzn5w\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823764 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823783 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5dgh7\" (UID: \"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823831 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64013929-bfea-444c-8bf9-afbc7625a690-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-th5xh\" (UID: \"64013929-bfea-444c-8bf9-afbc7625a690\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823849 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89ad9943-b0ac-4b88-aaf0-bdb332b22488-config\") pod \"kube-controller-manager-operator-78b949d7b-g42sk\" (UID: \"89ad9943-b0ac-4b88-aaf0-bdb332b22488\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823881 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823910 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-etcd-ca\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823930 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/89ad9943-b0ac-4b88-aaf0-bdb332b22488-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-g42sk\" (UID: \"89ad9943-b0ac-4b88-aaf0-bdb332b22488\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823957 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-r2xzk\" (UID: \"bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823975 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5dgh7\" (UID: \"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.823996 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824020 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-etcd-service-ca\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824042 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rchtx\" (UniqueName: \"kubernetes.io/projected/9d625f91-d4de-4687-bdbe-8ff985d7e406-kube-api-access-rchtx\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824073 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l62xf\" (UniqueName: \"kubernetes.io/projected/52b51ab7-5c1b-4cbb-a471-905a56f96a84-kube-api-access-l62xf\") pod \"dns-operator-744455d44c-w5lkb\" (UID: \"52b51ab7-5c1b-4cbb-a471-905a56f96a84\") " pod="openshift-dns-operator/dns-operator-744455d44c-w5lkb" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824095 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824118 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64013929-bfea-444c-8bf9-afbc7625a690-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-th5xh\" (UID: \"64013929-bfea-444c-8bf9-afbc7625a690\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824142 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824165 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-trusted-ca-bundle\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824187 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9x2w\" (UniqueName: \"kubernetes.io/projected/ebeadae0-aa3e-47d1-8068-1c339da55d91-kube-api-access-q9x2w\") pod \"machine-config-operator-74547568cd-6gqpg\" (UID: \"ebeadae0-aa3e-47d1-8068-1c339da55d91\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824209 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824243 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-serving-cert\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824262 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-config\") pod \"route-controller-manager-6576b87f9c-pzn5w\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824282 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5dgh7\" (UID: \"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824304 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d32d9b8b-2373-4c15-b5ce-f7f77c899cdd-srv-cert\") pod \"catalog-operator-68c6474976-6bfrv\" (UID: \"d32d9b8b-2373-4c15-b5ce-f7f77c899cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824324 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ebeadae0-aa3e-47d1-8068-1c339da55d91-proxy-tls\") pod \"machine-config-operator-74547568cd-6gqpg\" (UID: \"ebeadae0-aa3e-47d1-8068-1c339da55d91\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824355 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a111e46-9f4e-465f-821f-b58864174417-serving-cert\") pod \"openshift-config-operator-7777fb866f-mznfc\" (UID: \"5a111e46-9f4e-465f-821f-b58864174417\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824375 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-audit-policies\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824405 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-config\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824414 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/52b51ab7-5c1b-4cbb-a471-905a56f96a84-metrics-tls\") pod \"dns-operator-744455d44c-w5lkb\" (UID: \"52b51ab7-5c1b-4cbb-a471-905a56f96a84\") " pod="openshift-dns-operator/dns-operator-744455d44c-w5lkb" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824494 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824519 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824562 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-serving-cert\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824583 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-oauth-serving-cert\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824604 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5a111e46-9f4e-465f-821f-b58864174417-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mznfc\" (UID: \"5a111e46-9f4e-465f-821f-b58864174417\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824647 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824673 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kphgq\" (UniqueName: \"kubernetes.io/projected/586786e8-df0e-4424-8c84-bc2408b2dea5-kube-api-access-kphgq\") pod \"machine-approver-56656f9798-c4vmf\" (UID: \"586786e8-df0e-4424-8c84-bc2408b2dea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824715 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-client-ca\") pod \"route-controller-manager-6576b87f9c-pzn5w\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824746 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824813 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d32d9b8b-2373-4c15-b5ce-f7f77c899cdd-profile-collector-cert\") pod \"catalog-operator-68c6474976-6bfrv\" (UID: \"d32d9b8b-2373-4c15-b5ce-f7f77c899cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824842 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-serving-cert\") pod \"route-controller-manager-6576b87f9c-pzn5w\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.824902 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/596ad6ba-b5ec-471c-96b6-1b24d525b163-audit-dir\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.825187 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mznfc"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.825649 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/5a111e46-9f4e-465f-821f-b58864174417-available-featuregates\") pod \"openshift-config-operator-7777fb866f-mznfc\" (UID: \"5a111e46-9f4e-465f-821f-b58864174417\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.825900 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ebeadae0-aa3e-47d1-8068-1c339da55d91-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6gqpg\" (UID: \"ebeadae0-aa3e-47d1-8068-1c339da55d91\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.826203 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.826367 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/586786e8-df0e-4424-8c84-bc2408b2dea5-config\") pod \"machine-approver-56656f9798-c4vmf\" (UID: \"586786e8-df0e-4424-8c84-bc2408b2dea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.826424 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-cmb5q"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.826957 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/586786e8-df0e-4424-8c84-bc2408b2dea5-auth-proxy-config\") pod \"machine-approver-56656f9798-c4vmf\" (UID: \"586786e8-df0e-4424-8c84-bc2408b2dea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.828365 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-config\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.828944 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-etcd-service-ca\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.830004 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1-config\") pod \"kube-apiserver-operator-766d6c64bb-r2xzk\" (UID: \"bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.831731 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/586786e8-df0e-4424-8c84-bc2408b2dea5-machine-approver-tls\") pod \"machine-approver-56656f9798-c4vmf\" (UID: \"586786e8-df0e-4424-8c84-bc2408b2dea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.832377 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.832619 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-config\") pod \"route-controller-manager-6576b87f9c-pzn5w\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.833357 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.833919 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.833924 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89ad9943-b0ac-4b88-aaf0-bdb332b22488-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-g42sk\" (UID: \"89ad9943-b0ac-4b88-aaf0-bdb332b22488\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.834293 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.835322 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-client-ca\") pod \"route-controller-manager-6576b87f9c-pzn5w\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.835901 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/64013929-bfea-444c-8bf9-afbc7625a690-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-th5xh\" (UID: \"64013929-bfea-444c-8bf9-afbc7625a690\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.836059 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.836518 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-etcd-ca\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.837424 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-audit-policies\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.837583 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lw526"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.837625 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.838248 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89ad9943-b0ac-4b88-aaf0-bdb332b22488-config\") pod \"kube-controller-manager-operator-78b949d7b-g42sk\" (UID: \"89ad9943-b0ac-4b88-aaf0-bdb332b22488\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.838308 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-serving-cert\") pod \"route-controller-manager-6576b87f9c-pzn5w\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.838585 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.838811 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-r2xzk\" (UID: \"bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.838857 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/52b51ab7-5c1b-4cbb-a471-905a56f96a84-metrics-tls\") pod \"dns-operator-744455d44c-w5lkb\" (UID: \"52b51ab7-5c1b-4cbb-a471-905a56f96a84\") " pod="openshift-dns-operator/dns-operator-744455d44c-w5lkb" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.839240 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5dgh7\" (UID: \"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.839254 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.839838 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.840087 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-ldxc8"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.840361 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.841015 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-etcd-client\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.841908 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-zr6z9"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.842048 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ldxc8" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.842383 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-serving-cert\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.842664 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.843009 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64013929-bfea-444c-8bf9-afbc7625a690-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-th5xh\" (UID: \"64013929-bfea-444c-8bf9-afbc7625a690\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.845372 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a111e46-9f4e-465f-821f-b58864174417-serving-cert\") pod \"openshift-config-operator-7777fb866f-mznfc\" (UID: \"5a111e46-9f4e-465f-821f-b58864174417\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.845441 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.845568 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.846759 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.850327 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.851327 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.851408 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.852293 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-szpd6"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.853102 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-szpd6" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.855016 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.857425 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pplqc"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.858530 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-serving-cert\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.859224 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.863602 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.864938 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hx6pl"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.865977 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.867145 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-n7pfp"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.868407 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.870198 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.871538 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mk4bd"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.873321 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p975q"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.874728 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ztd4t"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.875677 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-oauth-config\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.876115 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.877204 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.878634 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-w5lkb"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.882069 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8k2xh"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.885993 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.886612 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.890338 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.890846 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.891964 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.893644 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lw526"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.894860 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.895915 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-service-ca\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.896412 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.897667 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.898914 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bsfdw"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.900203 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ntnp2"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.901714 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-szpd6"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.902883 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cmb5q"] Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.916185 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.923755 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-trusted-ca-bundle\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.930438 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.936931 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-oauth-serving-cert\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.950874 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.964079 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/d32d9b8b-2373-4c15-b5ce-f7f77c899cdd-profile-collector-cert\") pod \"catalog-operator-68c6474976-6bfrv\" (UID: \"d32d9b8b-2373-4c15-b5ce-f7f77c899cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.973567 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.991363 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 26 09:09:26 crc kubenswrapper[4872]: I0126 09:09:26.997213 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/d32d9b8b-2373-4c15-b5ce-f7f77c899cdd-srv-cert\") pod \"catalog-operator-68c6474976-6bfrv\" (UID: \"d32d9b8b-2373-4c15-b5ce-f7f77c899cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.010706 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.031197 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.050453 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.060681 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5dgh7\" (UID: \"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.075313 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.090180 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.097353 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ebeadae0-aa3e-47d1-8068-1c339da55d91-images\") pod \"machine-config-operator-74547568cd-6gqpg\" (UID: \"ebeadae0-aa3e-47d1-8068-1c339da55d91\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.110667 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.129612 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.139172 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ebeadae0-aa3e-47d1-8068-1c339da55d91-proxy-tls\") pod \"machine-config-operator-74547568cd-6gqpg\" (UID: \"ebeadae0-aa3e-47d1-8068-1c339da55d91\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.152350 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.171556 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.183899 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.183908 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.184730 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.190365 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.210862 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.231036 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.290258 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.310303 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.330751 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.351494 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.370603 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.391582 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.410685 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.431214 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.450374 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.470338 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.489746 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.525749 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msbxx\" (UniqueName: \"kubernetes.io/projected/b24744f1-cd08-4ac4-b6c5-13613cfa88d0-kube-api-access-msbxx\") pod \"apiserver-76f77b778f-82cml\" (UID: \"b24744f1-cd08-4ac4-b6c5-13613cfa88d0\") " pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.545544 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdspk\" (UniqueName: \"kubernetes.io/projected/f8486391-70e4-4b2f-a093-ea9ab42c18bd-kube-api-access-zdspk\") pod \"apiserver-7bbb656c7d-qglzk\" (UID: \"f8486391-70e4-4b2f-a093-ea9ab42c18bd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.566335 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrxkv\" (UniqueName: \"kubernetes.io/projected/07ffe425-ff10-4c9f-b3c5-fc2c5e167f67-kube-api-access-hrxkv\") pod \"machine-api-operator-5694c8668f-x9zlw\" (UID: \"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.571312 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.599854 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.610457 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.631688 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.651275 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.672313 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.692190 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.710663 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.728491 4872 request.go:700] Waited for 1.000693804s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/secrets?fieldSelector=metadata.name%3Dserving-cert&limit=500&resourceVersion=0 Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.730629 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.750462 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.768622 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz9xr\" (UniqueName: \"kubernetes.io/projected/3a3df101-99e3-41a0-a235-0876a7249434-kube-api-access-zz9xr\") pod \"controller-manager-879f6c89f-kxnj5\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.770780 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.785076 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.810501 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.811606 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.811985 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zscxs\" (UniqueName: \"kubernetes.io/projected/cde35d3c-8bd5-4e20-98de-2eca40d2a38c-kube-api-access-zscxs\") pod \"console-operator-58897d9998-zw7z6\" (UID: \"cde35d3c-8bd5-4e20-98de-2eca40d2a38c\") " pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.830923 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.853713 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.871331 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.891179 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.911474 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.930992 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.951362 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.972252 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.975438 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-x9zlw"] Jan 26 09:09:27 crc kubenswrapper[4872]: W0126 09:09:27.987547 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07ffe425_ff10_4c9f_b3c5_fc2c5e167f67.slice/crio-c97571f56983f4ce28b55ef057928fa385a5f0cba9203fa7936002fdacaf3514 WatchSource:0}: Error finding container c97571f56983f4ce28b55ef057928fa385a5f0cba9203fa7936002fdacaf3514: Status 404 returned error can't find the container with id c97571f56983f4ce28b55ef057928fa385a5f0cba9203fa7936002fdacaf3514 Jan 26 09:09:27 crc kubenswrapper[4872]: I0126 09:09:27.991319 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.011523 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.030956 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.031007 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.040645 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.051686 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.071476 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.091318 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.112004 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.131429 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.151258 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.172297 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.183488 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.192289 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.216050 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.231234 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.233201 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-zw7z6"] Jan 26 09:09:28 crc kubenswrapper[4872]: W0126 09:09:28.234380 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb24744f1_cd08_4ac4_b6c5_13613cfa88d0.slice/crio-37e996f52cf99ee8cb9d20988b21f2086fa0d4b5478161b372a7dc99b1a302ce WatchSource:0}: Error finding container 37e996f52cf99ee8cb9d20988b21f2086fa0d4b5478161b372a7dc99b1a302ce: Status 404 returned error can't find the container with id 37e996f52cf99ee8cb9d20988b21f2086fa0d4b5478161b372a7dc99b1a302ce Jan 26 09:09:28 crc kubenswrapper[4872]: W0126 09:09:28.236237 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcde35d3c_8bd5_4e20_98de_2eca40d2a38c.slice/crio-44ceeab4e420422a08c7d0e0b93c0ff0f5d1945ba3044e051d531839bcbbdc1c WatchSource:0}: Error finding container 44ceeab4e420422a08c7d0e0b93c0ff0f5d1945ba3044e051d531839bcbbdc1c: Status 404 returned error can't find the container with id 44ceeab4e420422a08c7d0e0b93c0ff0f5d1945ba3044e051d531839bcbbdc1c Jan 26 09:09:28 crc kubenswrapper[4872]: W0126 09:09:28.236997 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8486391_70e4_4b2f_a093_ea9ab42c18bd.slice/crio-4b0f6cd05593ded1125d5fc1747d88b7aea914089025cc4f1071143a5a7e65f7 WatchSource:0}: Error finding container 4b0f6cd05593ded1125d5fc1747d88b7aea914089025cc4f1071143a5a7e65f7: Status 404 returned error can't find the container with id 4b0f6cd05593ded1125d5fc1747d88b7aea914089025cc4f1071143a5a7e65f7 Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.237135 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-82cml"] Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.240965 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk"] Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.256215 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.268300 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kxnj5"] Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.270959 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.292999 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.311641 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.333630 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.358501 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.372413 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.391475 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.436946 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc98b\" (UniqueName: \"kubernetes.io/projected/d32d9b8b-2373-4c15-b5ce-f7f77c899cdd-kube-api-access-dc98b\") pod \"catalog-operator-68c6474976-6bfrv\" (UID: \"d32d9b8b-2373-4c15-b5ce-f7f77c899cdd\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.446884 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz6lv\" (UniqueName: \"kubernetes.io/projected/f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc-kube-api-access-lz6lv\") pod \"cluster-image-registry-operator-dc59b4c8b-5dgh7\" (UID: \"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.466700 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnj96\" (UniqueName: \"kubernetes.io/projected/340d9cdb-6984-4ccb-aaa3-f7a0e7402021-kube-api-access-gnj96\") pod \"etcd-operator-b45778765-hx6pl\" (UID: \"340d9cdb-6984-4ccb-aaa3-f7a0e7402021\") " pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.486960 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2cpv\" (UniqueName: \"kubernetes.io/projected/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-kube-api-access-g2cpv\") pod \"route-controller-manager-6576b87f9c-pzn5w\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.509058 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l62xf\" (UniqueName: \"kubernetes.io/projected/52b51ab7-5c1b-4cbb-a471-905a56f96a84-kube-api-access-l62xf\") pod \"dns-operator-744455d44c-w5lkb\" (UID: \"52b51ab7-5c1b-4cbb-a471-905a56f96a84\") " pod="openshift-dns-operator/dns-operator-744455d44c-w5lkb" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.524824 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.529632 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rchtx\" (UniqueName: \"kubernetes.io/projected/9d625f91-d4de-4687-bdbe-8ff985d7e406-kube-api-access-rchtx\") pod \"console-f9d7485db-ztd4t\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.546043 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrlrj\" (UniqueName: \"kubernetes.io/projected/5a111e46-9f4e-465f-821f-b58864174417-kube-api-access-rrlrj\") pod \"openshift-config-operator-7777fb866f-mznfc\" (UID: \"5a111e46-9f4e-465f-821f-b58864174417\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.565163 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8tzq\" (UniqueName: \"kubernetes.io/projected/596ad6ba-b5ec-471c-96b6-1b24d525b163-kube-api-access-l8tzq\") pod \"oauth-openshift-558db77b4-mk4bd\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.569159 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.587691 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqtpl\" (UniqueName: \"kubernetes.io/projected/64013929-bfea-444c-8bf9-afbc7625a690-kube-api-access-kqtpl\") pod \"openshift-controller-manager-operator-756b6f6bc6-th5xh\" (UID: \"64013929-bfea-444c-8bf9-afbc7625a690\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.590544 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.601226 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.610070 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9x2w\" (UniqueName: \"kubernetes.io/projected/ebeadae0-aa3e-47d1-8068-1c339da55d91-kube-api-access-q9x2w\") pod \"machine-config-operator-74547568cd-6gqpg\" (UID: \"ebeadae0-aa3e-47d1-8068-1c339da55d91\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.614621 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.614653 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.651213 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.651271 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5dgh7\" (UID: \"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.661381 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-w5lkb" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.671889 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.677124 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.692519 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.699324 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.716122 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/89ad9943-b0ac-4b88-aaf0-bdb332b22488-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-g42sk\" (UID: \"89ad9943-b0ac-4b88-aaf0-bdb332b22488\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.717236 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.727621 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w"] Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.729741 4872 request.go:700] Waited for 1.89295292s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-cluster-machine-approver/serviceaccounts/machine-approver-sa/token Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.737559 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-r2xzk\" (UID: \"bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.755604 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kphgq\" (UniqueName: \"kubernetes.io/projected/586786e8-df0e-4424-8c84-bc2408b2dea5-kube-api-access-kphgq\") pod \"machine-approver-56656f9798-c4vmf\" (UID: \"586786e8-df0e-4424-8c84-bc2408b2dea5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.757787 4872 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.776981 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.791416 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.813060 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.830723 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.851255 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.870877 4872 generic.go:334] "Generic (PLEG): container finished" podID="b24744f1-cd08-4ac4-b6c5-13613cfa88d0" containerID="8c36673477a885accf8343f6ab8e173d459148ec1cd8926a63aa65b92a0b6baa" exitCode=0 Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.870986 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-82cml" event={"ID":"b24744f1-cd08-4ac4-b6c5-13613cfa88d0","Type":"ContainerDied","Data":"8c36673477a885accf8343f6ab8e173d459148ec1cd8926a63aa65b92a0b6baa"} Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.871014 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-82cml" event={"ID":"b24744f1-cd08-4ac4-b6c5-13613cfa88d0","Type":"ContainerStarted","Data":"37e996f52cf99ee8cb9d20988b21f2086fa0d4b5478161b372a7dc99b1a302ce"} Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.875472 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.883015 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" event={"ID":"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8","Type":"ContainerStarted","Data":"ec8a4fe827f1a2aba42ae3f47a99412d92e051f12fd02c16644e68273fc5cb7d"} Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.886787 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-mznfc"] Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.900593 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.903827 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-zw7z6" event={"ID":"cde35d3c-8bd5-4e20-98de-2eca40d2a38c","Type":"ContainerStarted","Data":"982d087403b540240062873ed1fc15941066512779377cf3d20bd615fea9e88b"} Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.903874 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-zw7z6" event={"ID":"cde35d3c-8bd5-4e20-98de-2eca40d2a38c","Type":"ContainerStarted","Data":"44ceeab4e420422a08c7d0e0b93c0ff0f5d1945ba3044e051d531839bcbbdc1c"} Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.904108 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.910241 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" event={"ID":"3a3df101-99e3-41a0-a235-0876a7249434","Type":"ContainerStarted","Data":"c2f1f4116f3dbcafa5f89510171a630b400c108e120639da2f9e2fb3db9052aa"} Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.910304 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" event={"ID":"3a3df101-99e3-41a0-a235-0876a7249434","Type":"ContainerStarted","Data":"82ce8d27ada799de6b5b637d607ddc613ea37ca80e7163380080dd9d7a7006a6"} Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.910548 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.911764 4872 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-kxnj5 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.911834 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" podUID="3a3df101-99e3-41a0-a235-0876a7249434" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.912772 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.917991 4872 generic.go:334] "Generic (PLEG): container finished" podID="f8486391-70e4-4b2f-a093-ea9ab42c18bd" containerID="17678e9dceec462c2d333a43858fc2c91b862501b3698fe0fac8f5d33c75285b" exitCode=0 Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.918081 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" event={"ID":"f8486391-70e4-4b2f-a093-ea9ab42c18bd","Type":"ContainerDied","Data":"17678e9dceec462c2d333a43858fc2c91b862501b3698fe0fac8f5d33c75285b"} Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.918102 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" event={"ID":"f8486391-70e4-4b2f-a093-ea9ab42c18bd","Type":"ContainerStarted","Data":"4b0f6cd05593ded1125d5fc1747d88b7aea914089025cc4f1071143a5a7e65f7"} Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.921466 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" event={"ID":"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67","Type":"ContainerStarted","Data":"7457057858b24803c6445dc4840e4c700d36ae60c6effba19878549516a4f907"} Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.921529 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" event={"ID":"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67","Type":"ContainerStarted","Data":"781a1affeabf1545489bc71ca624f63d4a16fc57755e9a9c7b1ad5be2afba3e1"} Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.921544 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" event={"ID":"07ffe425-ff10-4c9f-b3c5-fc2c5e167f67","Type":"ContainerStarted","Data":"c97571f56983f4ce28b55ef057928fa385a5f0cba9203fa7936002fdacaf3514"} Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.931991 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.945350 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.950919 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.951339 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.971120 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 26 09:09:28 crc kubenswrapper[4872]: I0126 09:09:28.995939 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.012283 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.045264 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.057950 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-bound-sa-token\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058066 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/785b2c3d-cdc5-4c45-b4bb-3274cdb2a420-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nxv59\" (UID: \"785b2c3d-cdc5-4c45-b4bb-3274cdb2a420\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058114 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-registry-certificates\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058172 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058195 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/785b2c3d-cdc5-4c45-b4bb-3274cdb2a420-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nxv59\" (UID: \"785b2c3d-cdc5-4c45-b4bb-3274cdb2a420\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058257 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058285 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb2zr\" (UniqueName: \"kubernetes.io/projected/49b97b45-be77-455a-9a69-f20629af5940-kube-api-access-zb2zr\") pod \"cluster-samples-operator-665b6dd947-wfcsg\" (UID: \"49b97b45-be77-455a-9a69-f20629af5940\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058337 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpjq7\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-kube-api-access-xpjq7\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058362 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058392 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/49b97b45-be77-455a-9a69-f20629af5940-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wfcsg\" (UID: \"49b97b45-be77-455a-9a69-f20629af5940\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058433 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-trusted-ca\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058465 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-registry-tls\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058487 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l7bk\" (UniqueName: \"kubernetes.io/projected/785b2c3d-cdc5-4c45-b4bb-3274cdb2a420-kube-api-access-2l7bk\") pod \"openshift-apiserver-operator-796bbdcf4f-nxv59\" (UID: \"785b2c3d-cdc5-4c45-b4bb-3274cdb2a420\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.058515 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65vnh\" (UniqueName: \"kubernetes.io/projected/c85d43e4-21f0-47fc-ae62-057de42e1935-kube-api-access-65vnh\") pod \"downloads-7954f5f757-zr6z9\" (UID: \"c85d43e4-21f0-47fc-ae62-057de42e1935\") " pod="openshift-console/downloads-7954f5f757-zr6z9" Jan 26 09:09:29 crc kubenswrapper[4872]: E0126 09:09:29.059204 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:29.559187332 +0000 UTC m=+102.868027133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.071272 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.094489 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 26 09:09:29 crc kubenswrapper[4872]: W0126 09:09:29.105196 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod586786e8_df0e_4424_8c84_bc2408b2dea5.slice/crio-14d620988af2e378d67b4be6e2a3243c87686a3f218f37594b57e559fff2434b WatchSource:0}: Error finding container 14d620988af2e378d67b4be6e2a3243c87686a3f218f37594b57e559fff2434b: Status 404 returned error can't find the container with id 14d620988af2e378d67b4be6e2a3243c87686a3f218f37594b57e559fff2434b Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.159881 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:29 crc kubenswrapper[4872]: E0126 09:09:29.160328 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:29.660282069 +0000 UTC m=+102.969121870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161023 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/835cb68e-1243-4522-bfd0-181378353e51-bound-sa-token\") pod \"ingress-operator-5b745b69d9-nwkxd\" (UID: \"835cb68e-1243-4522-bfd0-181378353e51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161159 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2c67438-c0c7-48b8-a3df-21669ccb1602-secret-volume\") pod \"collect-profiles-29490300-2v9qs\" (UID: \"f2c67438-c0c7-48b8-a3df-21669ccb1602\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161218 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-trusted-ca\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161240 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/031837e9-7dec-4179-9194-3b95092a8569-serving-cert\") pod \"service-ca-operator-777779d784-vvzw4\" (UID: \"031837e9-7dec-4179-9194-3b95092a8569\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161304 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-default-certificate\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161534 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-stats-auth\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161622 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-registry-tls\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161652 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a8658e2c-7305-4770-9b0d-1097def3df9f-signing-key\") pod \"service-ca-9c57cc56f-n7pfp\" (UID: \"a8658e2c-7305-4770-9b0d-1097def3df9f\") " pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161725 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65vnh\" (UniqueName: \"kubernetes.io/projected/c85d43e4-21f0-47fc-ae62-057de42e1935-kube-api-access-65vnh\") pod \"downloads-7954f5f757-zr6z9\" (UID: \"c85d43e4-21f0-47fc-ae62-057de42e1935\") " pod="openshift-console/downloads-7954f5f757-zr6z9" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161786 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2c67438-c0c7-48b8-a3df-21669ccb1602-config-volume\") pod \"collect-profiles-29490300-2v9qs\" (UID: \"f2c67438-c0c7-48b8-a3df-21669ccb1602\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161841 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-registration-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161912 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/785b2c3d-cdc5-4c45-b4bb-3274cdb2a420-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nxv59\" (UID: \"785b2c3d-cdc5-4c45-b4bb-3274cdb2a420\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161942 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b1a66ff-c739-4077-ab76-bf675e158a77-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hvnl5\" (UID: \"9b1a66ff-c739-4077-ab76-bf675e158a77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.161990 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7e3a84de-b4dd-4fc0-a279-9e2a9a616216-metrics-tls\") pod \"dns-default-cmb5q\" (UID: \"7e3a84de-b4dd-4fc0-a279-9e2a9a616216\") " pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162061 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-bound-sa-token\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162152 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjd9h\" (UniqueName: \"kubernetes.io/projected/238152aa-010b-437e-8439-f8197970d5f9-kube-api-access-xjd9h\") pod \"olm-operator-6b444d44fb-gfcvb\" (UID: \"238152aa-010b-437e-8439-f8197970d5f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162230 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf9fk\" (UniqueName: \"kubernetes.io/projected/437f0c85-bf20-4e24-9138-cce77c62fd2c-kube-api-access-tf9fk\") pod \"multus-admission-controller-857f4d67dd-8k2xh\" (UID: \"437f0c85-bf20-4e24-9138-cce77c62fd2c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8k2xh" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162306 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031837e9-7dec-4179-9194-3b95092a8569-config\") pod \"service-ca-operator-777779d784-vvzw4\" (UID: \"031837e9-7dec-4179-9194-3b95092a8569\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162331 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a8658e2c-7305-4770-9b0d-1097def3df9f-signing-cabundle\") pod \"service-ca-9c57cc56f-n7pfp\" (UID: \"a8658e2c-7305-4770-9b0d-1097def3df9f\") " pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162354 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7975e724-e9f6-4a68-8779-f894ac687f74-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162472 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/835cb68e-1243-4522-bfd0-181378353e51-metrics-tls\") pod \"ingress-operator-5b745b69d9-nwkxd\" (UID: \"835cb68e-1243-4522-bfd0-181378353e51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162499 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e3a84de-b4dd-4fc0-a279-9e2a9a616216-config-volume\") pod \"dns-default-cmb5q\" (UID: \"7e3a84de-b4dd-4fc0-a279-9e2a9a616216\") " pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162537 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/785b2c3d-cdc5-4c45-b4bb-3274cdb2a420-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nxv59\" (UID: \"785b2c3d-cdc5-4c45-b4bb-3274cdb2a420\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162566 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvbm2\" (UniqueName: \"kubernetes.io/projected/5473931a-2e46-45af-81f5-89fe0d299e4b-kube-api-access-tvbm2\") pod \"machine-config-controller-84d6567774-qrfsk\" (UID: \"5473931a-2e46-45af-81f5-89fe0d299e4b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162604 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162632 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb2zr\" (UniqueName: \"kubernetes.io/projected/49b97b45-be77-455a-9a69-f20629af5940-kube-api-access-zb2zr\") pod \"cluster-samples-operator-665b6dd947-wfcsg\" (UID: \"49b97b45-be77-455a-9a69-f20629af5940\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162664 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-plugins-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162783 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tkfq\" (UniqueName: \"kubernetes.io/projected/7e3a84de-b4dd-4fc0-a279-9e2a9a616216-kube-api-access-6tkfq\") pod \"dns-default-cmb5q\" (UID: \"7e3a84de-b4dd-4fc0-a279-9e2a9a616216\") " pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162883 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7975e724-e9f6-4a68-8779-f894ac687f74-service-ca-bundle\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162913 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f3e58300-1dc1-4f00-a78d-e33802c262f1-node-bootstrap-token\") pod \"machine-config-server-ldxc8\" (UID: \"f3e58300-1dc1-4f00-a78d-e33802c262f1\") " pod="openshift-machine-config-operator/machine-config-server-ldxc8" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162938 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5473931a-2e46-45af-81f5-89fe0d299e4b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qrfsk\" (UID: \"5473931a-2e46-45af-81f5-89fe0d299e4b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.162993 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7n6b\" (UniqueName: \"kubernetes.io/projected/1bd8496f-6499-4cf1-894a-37f68df536f4-kube-api-access-v7n6b\") pod \"migrator-59844c95c7-pplqc\" (UID: \"1bd8496f-6499-4cf1-894a-37f68df536f4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pplqc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163030 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpjq7\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-kube-api-access-xpjq7\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163089 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zdlw\" (UniqueName: \"kubernetes.io/projected/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-kube-api-access-7zdlw\") pod \"marketplace-operator-79b997595-bsfdw\" (UID: \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163135 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b8096afa-b0ad-4a6d-b301-108d746d6bbb-tmpfs\") pod \"packageserver-d55dfcdfc-5z7tb\" (UID: \"b8096afa-b0ad-4a6d-b301-108d746d6bbb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163190 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163219 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-service-ca-bundle\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163291 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcr9k\" (UniqueName: \"kubernetes.io/projected/b8096afa-b0ad-4a6d-b301-108d746d6bbb-kube-api-access-rcr9k\") pod \"packageserver-d55dfcdfc-5z7tb\" (UID: \"b8096afa-b0ad-4a6d-b301-108d746d6bbb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163322 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l75vq\" (UniqueName: \"kubernetes.io/projected/835cb68e-1243-4522-bfd0-181378353e51-kube-api-access-l75vq\") pod \"ingress-operator-5b745b69d9-nwkxd\" (UID: \"835cb68e-1243-4522-bfd0-181378353e51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163350 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4hhdc\" (UID: \"ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163391 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8sfc\" (UniqueName: \"kubernetes.io/projected/f3e58300-1dc1-4f00-a78d-e33802c262f1-kube-api-access-x8sfc\") pod \"machine-config-server-ldxc8\" (UID: \"f3e58300-1dc1-4f00-a78d-e33802c262f1\") " pod="openshift-machine-config-operator/machine-config-server-ldxc8" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163418 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpthw\" (UniqueName: \"kubernetes.io/projected/a155afc3-d969-48f2-abc6-720c1f036e3c-kube-api-access-xpthw\") pod \"ingress-canary-szpd6\" (UID: \"a155afc3-d969-48f2-abc6-720c1f036e3c\") " pod="openshift-ingress-canary/ingress-canary-szpd6" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163442 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4hhdc\" (UID: \"ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163491 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkpjv\" (UniqueName: \"kubernetes.io/projected/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-kube-api-access-lkpjv\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163535 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7bsh\" (UniqueName: \"kubernetes.io/projected/9b1a66ff-c739-4077-ab76-bf675e158a77-kube-api-access-l7bsh\") pod \"control-plane-machine-set-operator-78cbb6b69f-hvnl5\" (UID: \"9b1a66ff-c739-4077-ab76-bf675e158a77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163573 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqr7m\" (UniqueName: \"kubernetes.io/projected/7975e724-e9f6-4a68-8779-f894ac687f74-kube-api-access-lqr7m\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163600 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vpt9\" (UniqueName: \"kubernetes.io/projected/f2c67438-c0c7-48b8-a3df-21669ccb1602-kube-api-access-8vpt9\") pod \"collect-profiles-29490300-2v9qs\" (UID: \"f2c67438-c0c7-48b8-a3df-21669ccb1602\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163625 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck78b\" (UniqueName: \"kubernetes.io/projected/ea814b88-1c4c-407b-8754-f57e19f8153f-kube-api-access-ck78b\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163661 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j98s\" (UniqueName: \"kubernetes.io/projected/a8658e2c-7305-4770-9b0d-1097def3df9f-kube-api-access-5j98s\") pod \"service-ca-9c57cc56f-n7pfp\" (UID: \"a8658e2c-7305-4770-9b0d-1097def3df9f\") " pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163683 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/835cb68e-1243-4522-bfd0-181378353e51-trusted-ca\") pod \"ingress-operator-5b745b69d9-nwkxd\" (UID: \"835cb68e-1243-4522-bfd0-181378353e51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163706 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87dwr\" (UniqueName: \"kubernetes.io/projected/06e81723-7f8b-4dd6-99c6-d1ce1a966b7d-kube-api-access-87dwr\") pod \"kube-storage-version-migrator-operator-b67b599dd-zbn9p\" (UID: \"06e81723-7f8b-4dd6-99c6-d1ce1a966b7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163745 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l7bk\" (UniqueName: \"kubernetes.io/projected/785b2c3d-cdc5-4c45-b4bb-3274cdb2a420-kube-api-access-2l7bk\") pod \"openshift-apiserver-operator-796bbdcf4f-nxv59\" (UID: \"785b2c3d-cdc5-4c45-b4bb-3274cdb2a420\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163767 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/55d2e904-b5e7-4bcf-8c39-1e84e51a2581-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-dcs6q\" (UID: \"55d2e904-b5e7-4bcf-8c39-1e84e51a2581\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163792 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-socket-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163890 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7975e724-e9f6-4a68-8779-f894ac687f74-serving-cert\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163928 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/238152aa-010b-437e-8439-f8197970d5f9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gfcvb\" (UID: \"238152aa-010b-437e-8439-f8197970d5f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163950 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7975e724-e9f6-4a68-8779-f894ac687f74-config\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.163987 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-metrics-certs\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.164023 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd54l\" (UniqueName: \"kubernetes.io/projected/55d2e904-b5e7-4bcf-8c39-1e84e51a2581-kube-api-access-xd54l\") pod \"package-server-manager-789f6589d5-dcs6q\" (UID: \"55d2e904-b5e7-4bcf-8c39-1e84e51a2581\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.164043 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-mountpoint-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.164062 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/437f0c85-bf20-4e24-9138-cce77c62fd2c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8k2xh\" (UID: \"437f0c85-bf20-4e24-9138-cce77c62fd2c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8k2xh" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.164082 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4hhdc\" (UID: \"ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.164120 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-registry-certificates\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.164153 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-csi-data-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.164188 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b8096afa-b0ad-4a6d-b301-108d746d6bbb-webhook-cert\") pod \"packageserver-d55dfcdfc-5z7tb\" (UID: \"b8096afa-b0ad-4a6d-b301-108d746d6bbb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.164317 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.164581 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bsfdw\" (UID: \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.164615 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f3e58300-1dc1-4f00-a78d-e33802c262f1-certs\") pod \"machine-config-server-ldxc8\" (UID: \"f3e58300-1dc1-4f00-a78d-e33802c262f1\") " pod="openshift-machine-config-operator/machine-config-server-ldxc8" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.165019 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5hh4\" (UniqueName: \"kubernetes.io/projected/031837e9-7dec-4179-9194-3b95092a8569-kube-api-access-v5hh4\") pod \"service-ca-operator-777779d784-vvzw4\" (UID: \"031837e9-7dec-4179-9194-3b95092a8569\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.165084 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e81723-7f8b-4dd6-99c6-d1ce1a966b7d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-zbn9p\" (UID: \"06e81723-7f8b-4dd6-99c6-d1ce1a966b7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.165108 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a155afc3-d969-48f2-abc6-720c1f036e3c-cert\") pod \"ingress-canary-szpd6\" (UID: \"a155afc3-d969-48f2-abc6-720c1f036e3c\") " pod="openshift-ingress-canary/ingress-canary-szpd6" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.165144 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b8096afa-b0ad-4a6d-b301-108d746d6bbb-apiservice-cert\") pod \"packageserver-d55dfcdfc-5z7tb\" (UID: \"b8096afa-b0ad-4a6d-b301-108d746d6bbb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.165183 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/238152aa-010b-437e-8439-f8197970d5f9-srv-cert\") pod \"olm-operator-6b444d44fb-gfcvb\" (UID: \"238152aa-010b-437e-8439-f8197970d5f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.165204 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bsfdw\" (UID: \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.165229 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e81723-7f8b-4dd6-99c6-d1ce1a966b7d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-zbn9p\" (UID: \"06e81723-7f8b-4dd6-99c6-d1ce1a966b7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.165249 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5473931a-2e46-45af-81f5-89fe0d299e4b-proxy-tls\") pod \"machine-config-controller-84d6567774-qrfsk\" (UID: \"5473931a-2e46-45af-81f5-89fe0d299e4b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.165288 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/49b97b45-be77-455a-9a69-f20629af5940-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wfcsg\" (UID: \"49b97b45-be77-455a-9a69-f20629af5940\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.215226 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/785b2c3d-cdc5-4c45-b4bb-3274cdb2a420-config\") pod \"openshift-apiserver-operator-796bbdcf4f-nxv59\" (UID: \"785b2c3d-cdc5-4c45-b4bb-3274cdb2a420\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.234206 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-trusted-ca\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.292905 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65vnh\" (UniqueName: \"kubernetes.io/projected/c85d43e4-21f0-47fc-ae62-057de42e1935-kube-api-access-65vnh\") pod \"downloads-7954f5f757-zr6z9\" (UID: \"c85d43e4-21f0-47fc-ae62-057de42e1935\") " pod="openshift-console/downloads-7954f5f757-zr6z9" Jan 26 09:09:29 crc kubenswrapper[4872]: E0126 09:09:29.295787 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:29.795727399 +0000 UTC m=+103.104567200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.296145 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.296477 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e81723-7f8b-4dd6-99c6-d1ce1a966b7d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-zbn9p\" (UID: \"06e81723-7f8b-4dd6-99c6-d1ce1a966b7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.296511 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5473931a-2e46-45af-81f5-89fe0d299e4b-proxy-tls\") pod \"machine-config-controller-84d6567774-qrfsk\" (UID: \"5473931a-2e46-45af-81f5-89fe0d299e4b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.296553 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/835cb68e-1243-4522-bfd0-181378353e51-bound-sa-token\") pod \"ingress-operator-5b745b69d9-nwkxd\" (UID: \"835cb68e-1243-4522-bfd0-181378353e51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.296576 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/031837e9-7dec-4179-9194-3b95092a8569-serving-cert\") pod \"service-ca-operator-777779d784-vvzw4\" (UID: \"031837e9-7dec-4179-9194-3b95092a8569\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.296594 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2c67438-c0c7-48b8-a3df-21669ccb1602-secret-volume\") pod \"collect-profiles-29490300-2v9qs\" (UID: \"f2c67438-c0c7-48b8-a3df-21669ccb1602\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.296617 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-default-certificate\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297278 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-stats-auth\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297309 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a8658e2c-7305-4770-9b0d-1097def3df9f-signing-key\") pod \"service-ca-9c57cc56f-n7pfp\" (UID: \"a8658e2c-7305-4770-9b0d-1097def3df9f\") " pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297465 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2c67438-c0c7-48b8-a3df-21669ccb1602-config-volume\") pod \"collect-profiles-29490300-2v9qs\" (UID: \"f2c67438-c0c7-48b8-a3df-21669ccb1602\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297490 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-registration-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297540 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7e3a84de-b4dd-4fc0-a279-9e2a9a616216-metrics-tls\") pod \"dns-default-cmb5q\" (UID: \"7e3a84de-b4dd-4fc0-a279-9e2a9a616216\") " pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297576 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b1a66ff-c739-4077-ab76-bf675e158a77-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hvnl5\" (UID: \"9b1a66ff-c739-4077-ab76-bf675e158a77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297626 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjd9h\" (UniqueName: \"kubernetes.io/projected/238152aa-010b-437e-8439-f8197970d5f9-kube-api-access-xjd9h\") pod \"olm-operator-6b444d44fb-gfcvb\" (UID: \"238152aa-010b-437e-8439-f8197970d5f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297732 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf9fk\" (UniqueName: \"kubernetes.io/projected/437f0c85-bf20-4e24-9138-cce77c62fd2c-kube-api-access-tf9fk\") pod \"multus-admission-controller-857f4d67dd-8k2xh\" (UID: \"437f0c85-bf20-4e24-9138-cce77c62fd2c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8k2xh" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297759 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031837e9-7dec-4179-9194-3b95092a8569-config\") pod \"service-ca-operator-777779d784-vvzw4\" (UID: \"031837e9-7dec-4179-9194-3b95092a8569\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297777 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a8658e2c-7305-4770-9b0d-1097def3df9f-signing-cabundle\") pod \"service-ca-9c57cc56f-n7pfp\" (UID: \"a8658e2c-7305-4770-9b0d-1097def3df9f\") " pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297852 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7975e724-e9f6-4a68-8779-f894ac687f74-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297930 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/835cb68e-1243-4522-bfd0-181378353e51-metrics-tls\") pod \"ingress-operator-5b745b69d9-nwkxd\" (UID: \"835cb68e-1243-4522-bfd0-181378353e51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.297958 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e3a84de-b4dd-4fc0-a279-9e2a9a616216-config-volume\") pod \"dns-default-cmb5q\" (UID: \"7e3a84de-b4dd-4fc0-a279-9e2a9a616216\") " pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298002 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvbm2\" (UniqueName: \"kubernetes.io/projected/5473931a-2e46-45af-81f5-89fe0d299e4b-kube-api-access-tvbm2\") pod \"machine-config-controller-84d6567774-qrfsk\" (UID: \"5473931a-2e46-45af-81f5-89fe0d299e4b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298064 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-plugins-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298135 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tkfq\" (UniqueName: \"kubernetes.io/projected/7e3a84de-b4dd-4fc0-a279-9e2a9a616216-kube-api-access-6tkfq\") pod \"dns-default-cmb5q\" (UID: \"7e3a84de-b4dd-4fc0-a279-9e2a9a616216\") " pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298188 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7975e724-e9f6-4a68-8779-f894ac687f74-service-ca-bundle\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298223 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5473931a-2e46-45af-81f5-89fe0d299e4b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qrfsk\" (UID: \"5473931a-2e46-45af-81f5-89fe0d299e4b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298252 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f3e58300-1dc1-4f00-a78d-e33802c262f1-node-bootstrap-token\") pod \"machine-config-server-ldxc8\" (UID: \"f3e58300-1dc1-4f00-a78d-e33802c262f1\") " pod="openshift-machine-config-operator/machine-config-server-ldxc8" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298283 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7n6b\" (UniqueName: \"kubernetes.io/projected/1bd8496f-6499-4cf1-894a-37f68df536f4-kube-api-access-v7n6b\") pod \"migrator-59844c95c7-pplqc\" (UID: \"1bd8496f-6499-4cf1-894a-37f68df536f4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pplqc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298330 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zdlw\" (UniqueName: \"kubernetes.io/projected/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-kube-api-access-7zdlw\") pod \"marketplace-operator-79b997595-bsfdw\" (UID: \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298358 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b8096afa-b0ad-4a6d-b301-108d746d6bbb-tmpfs\") pod \"packageserver-d55dfcdfc-5z7tb\" (UID: \"b8096afa-b0ad-4a6d-b301-108d746d6bbb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298399 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-service-ca-bundle\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298450 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l75vq\" (UniqueName: \"kubernetes.io/projected/835cb68e-1243-4522-bfd0-181378353e51-kube-api-access-l75vq\") pod \"ingress-operator-5b745b69d9-nwkxd\" (UID: \"835cb68e-1243-4522-bfd0-181378353e51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298485 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcr9k\" (UniqueName: \"kubernetes.io/projected/b8096afa-b0ad-4a6d-b301-108d746d6bbb-kube-api-access-rcr9k\") pod \"packageserver-d55dfcdfc-5z7tb\" (UID: \"b8096afa-b0ad-4a6d-b301-108d746d6bbb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298514 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4hhdc\" (UID: \"ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298567 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpthw\" (UniqueName: \"kubernetes.io/projected/a155afc3-d969-48f2-abc6-720c1f036e3c-kube-api-access-xpthw\") pod \"ingress-canary-szpd6\" (UID: \"a155afc3-d969-48f2-abc6-720c1f036e3c\") " pod="openshift-ingress-canary/ingress-canary-szpd6" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298595 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4hhdc\" (UID: \"ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298650 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8sfc\" (UniqueName: \"kubernetes.io/projected/f3e58300-1dc1-4f00-a78d-e33802c262f1-kube-api-access-x8sfc\") pod \"machine-config-server-ldxc8\" (UID: \"f3e58300-1dc1-4f00-a78d-e33802c262f1\") " pod="openshift-machine-config-operator/machine-config-server-ldxc8" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298679 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqr7m\" (UniqueName: \"kubernetes.io/projected/7975e724-e9f6-4a68-8779-f894ac687f74-kube-api-access-lqr7m\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298703 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/a8658e2c-7305-4770-9b0d-1097def3df9f-signing-cabundle\") pod \"service-ca-9c57cc56f-n7pfp\" (UID: \"a8658e2c-7305-4770-9b0d-1097def3df9f\") " pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298705 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkpjv\" (UniqueName: \"kubernetes.io/projected/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-kube-api-access-lkpjv\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298764 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7bsh\" (UniqueName: \"kubernetes.io/projected/9b1a66ff-c739-4077-ab76-bf675e158a77-kube-api-access-l7bsh\") pod \"control-plane-machine-set-operator-78cbb6b69f-hvnl5\" (UID: \"9b1a66ff-c739-4077-ab76-bf675e158a77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298792 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j98s\" (UniqueName: \"kubernetes.io/projected/a8658e2c-7305-4770-9b0d-1097def3df9f-kube-api-access-5j98s\") pod \"service-ca-9c57cc56f-n7pfp\" (UID: \"a8658e2c-7305-4770-9b0d-1097def3df9f\") " pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298835 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/835cb68e-1243-4522-bfd0-181378353e51-trusted-ca\") pod \"ingress-operator-5b745b69d9-nwkxd\" (UID: \"835cb68e-1243-4522-bfd0-181378353e51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298894 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vpt9\" (UniqueName: \"kubernetes.io/projected/f2c67438-c0c7-48b8-a3df-21669ccb1602-kube-api-access-8vpt9\") pod \"collect-profiles-29490300-2v9qs\" (UID: \"f2c67438-c0c7-48b8-a3df-21669ccb1602\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298918 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck78b\" (UniqueName: \"kubernetes.io/projected/ea814b88-1c4c-407b-8754-f57e19f8153f-kube-api-access-ck78b\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.298972 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/55d2e904-b5e7-4bcf-8c39-1e84e51a2581-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-dcs6q\" (UID: \"55d2e904-b5e7-4bcf-8c39-1e84e51a2581\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299044 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87dwr\" (UniqueName: \"kubernetes.io/projected/06e81723-7f8b-4dd6-99c6-d1ce1a966b7d-kube-api-access-87dwr\") pod \"kube-storage-version-migrator-operator-b67b599dd-zbn9p\" (UID: \"06e81723-7f8b-4dd6-99c6-d1ce1a966b7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299074 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-socket-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299096 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7975e724-e9f6-4a68-8779-f894ac687f74-serving-cert\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299119 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/238152aa-010b-437e-8439-f8197970d5f9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gfcvb\" (UID: \"238152aa-010b-437e-8439-f8197970d5f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299140 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7975e724-e9f6-4a68-8779-f894ac687f74-config\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299164 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-metrics-certs\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299191 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4hhdc\" (UID: \"ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299223 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd54l\" (UniqueName: \"kubernetes.io/projected/55d2e904-b5e7-4bcf-8c39-1e84e51a2581-kube-api-access-xd54l\") pod \"package-server-manager-789f6589d5-dcs6q\" (UID: \"55d2e904-b5e7-4bcf-8c39-1e84e51a2581\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299244 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-mountpoint-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299280 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-csi-data-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299303 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b8096afa-b0ad-4a6d-b301-108d746d6bbb-webhook-cert\") pod \"packageserver-d55dfcdfc-5z7tb\" (UID: \"b8096afa-b0ad-4a6d-b301-108d746d6bbb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299376 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f3e58300-1dc1-4f00-a78d-e33802c262f1-certs\") pod \"machine-config-server-ldxc8\" (UID: \"f3e58300-1dc1-4f00-a78d-e33802c262f1\") " pod="openshift-machine-config-operator/machine-config-server-ldxc8" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299400 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bsfdw\" (UID: \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299432 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5hh4\" (UniqueName: \"kubernetes.io/projected/031837e9-7dec-4179-9194-3b95092a8569-kube-api-access-v5hh4\") pod \"service-ca-operator-777779d784-vvzw4\" (UID: \"031837e9-7dec-4179-9194-3b95092a8569\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299454 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a155afc3-d969-48f2-abc6-720c1f036e3c-cert\") pod \"ingress-canary-szpd6\" (UID: \"a155afc3-d969-48f2-abc6-720c1f036e3c\") " pod="openshift-ingress-canary/ingress-canary-szpd6" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299473 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e81723-7f8b-4dd6-99c6-d1ce1a966b7d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-zbn9p\" (UID: \"06e81723-7f8b-4dd6-99c6-d1ce1a966b7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299501 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b8096afa-b0ad-4a6d-b301-108d746d6bbb-apiservice-cert\") pod \"packageserver-d55dfcdfc-5z7tb\" (UID: \"b8096afa-b0ad-4a6d-b301-108d746d6bbb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299522 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/238152aa-010b-437e-8439-f8197970d5f9-srv-cert\") pod \"olm-operator-6b444d44fb-gfcvb\" (UID: \"238152aa-010b-437e-8439-f8197970d5f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299546 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bsfdw\" (UID: \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.300230 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7975e724-e9f6-4a68-8779-f894ac687f74-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.300461 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7975e724-e9f6-4a68-8779-f894ac687f74-service-ca-bundle\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.300679 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b8096afa-b0ad-4a6d-b301-108d746d6bbb-tmpfs\") pod \"packageserver-d55dfcdfc-5z7tb\" (UID: \"b8096afa-b0ad-4a6d-b301-108d746d6bbb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.301496 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5473931a-2e46-45af-81f5-89fe0d299e4b-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-qrfsk\" (UID: \"5473931a-2e46-45af-81f5-89fe0d299e4b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.305855 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e3a84de-b4dd-4fc0-a279-9e2a9a616216-config-volume\") pod \"dns-default-cmb5q\" (UID: \"7e3a84de-b4dd-4fc0-a279-9e2a9a616216\") " pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.306753 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-plugins-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.307074 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bsfdw\" (UID: \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.307167 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/49b97b45-be77-455a-9a69-f20629af5940-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wfcsg\" (UID: \"49b97b45-be77-455a-9a69-f20629af5940\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.307454 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-registry-tls\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.308229 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06e81723-7f8b-4dd6-99c6-d1ce1a966b7d-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-zbn9p\" (UID: \"06e81723-7f8b-4dd6-99c6-d1ce1a966b7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.308460 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-service-ca-bundle\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.310382 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/785b2c3d-cdc5-4c45-b4bb-3274cdb2a420-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-nxv59\" (UID: \"785b2c3d-cdc5-4c45-b4bb-3274cdb2a420\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.299237 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-registry-certificates\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.316066 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-zw7z6" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.316100 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7"] Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.316115 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-hx6pl"] Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.316058 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f3e58300-1dc1-4f00-a78d-e33802c262f1-node-bootstrap-token\") pod \"machine-config-server-ldxc8\" (UID: \"f3e58300-1dc1-4f00-a78d-e33802c262f1\") " pod="openshift-machine-config-operator/machine-config-server-ldxc8" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.317894 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb2zr\" (UniqueName: \"kubernetes.io/projected/49b97b45-be77-455a-9a69-f20629af5940-kube-api-access-zb2zr\") pod \"cluster-samples-operator-665b6dd947-wfcsg\" (UID: \"49b97b45-be77-455a-9a69-f20629af5940\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.319960 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/835cb68e-1243-4522-bfd0-181378353e51-metrics-tls\") pod \"ingress-operator-5b745b69d9-nwkxd\" (UID: \"835cb68e-1243-4522-bfd0-181378353e51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.326231 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7975e724-e9f6-4a68-8779-f894ac687f74-config\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.326618 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-socket-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.329990 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-csi-data-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.330347 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5473931a-2e46-45af-81f5-89fe0d299e4b-proxy-tls\") pod \"machine-config-controller-84d6567774-qrfsk\" (UID: \"5473931a-2e46-45af-81f5-89fe0d299e4b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.333594 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh"] Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.333836 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-mountpoint-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.335016 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.335073 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4hhdc\" (UID: \"ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.337182 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mk4bd"] Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.342326 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2c67438-c0c7-48b8-a3df-21669ccb1602-config-volume\") pod \"collect-profiles-29490300-2v9qs\" (UID: \"f2c67438-c0c7-48b8-a3df-21669ccb1602\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.342675 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ea814b88-1c4c-407b-8754-f57e19f8153f-registration-dir\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.346159 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031837e9-7dec-4179-9194-3b95092a8569-config\") pod \"service-ca-operator-777779d784-vvzw4\" (UID: \"031837e9-7dec-4179-9194-3b95092a8569\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.346629 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zdlw\" (UniqueName: \"kubernetes.io/projected/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-kube-api-access-7zdlw\") pod \"marketplace-operator-79b997595-bsfdw\" (UID: \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.353112 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkpjv\" (UniqueName: \"kubernetes.io/projected/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-kube-api-access-lkpjv\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.363270 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/a8658e2c-7305-4770-9b0d-1097def3df9f-signing-key\") pod \"service-ca-9c57cc56f-n7pfp\" (UID: \"a8658e2c-7305-4770-9b0d-1097def3df9f\") " pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" Jan 26 09:09:29 crc kubenswrapper[4872]: E0126 09:09:29.364955 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:29.86493317 +0000 UTC m=+103.173772971 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.365696 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.368691 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b8096afa-b0ad-4a6d-b301-108d746d6bbb-apiservice-cert\") pod \"packageserver-d55dfcdfc-5z7tb\" (UID: \"b8096afa-b0ad-4a6d-b301-108d746d6bbb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.371148 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06e81723-7f8b-4dd6-99c6-d1ce1a966b7d-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-zbn9p\" (UID: \"06e81723-7f8b-4dd6-99c6-d1ce1a966b7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.371159 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2c67438-c0c7-48b8-a3df-21669ccb1602-secret-volume\") pod \"collect-profiles-29490300-2v9qs\" (UID: \"f2c67438-c0c7-48b8-a3df-21669ccb1602\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.372763 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/55d2e904-b5e7-4bcf-8c39-1e84e51a2581-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-dcs6q\" (UID: \"55d2e904-b5e7-4bcf-8c39-1e84e51a2581\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.375304 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b1a66ff-c739-4077-ab76-bf675e158a77-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-hvnl5\" (UID: \"9b1a66ff-c739-4077-ab76-bf675e158a77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.376394 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bsfdw\" (UID: \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\") " pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.379647 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7e3a84de-b4dd-4fc0-a279-9e2a9a616216-metrics-tls\") pod \"dns-default-cmb5q\" (UID: \"7e3a84de-b4dd-4fc0-a279-9e2a9a616216\") " pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.380083 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f3e58300-1dc1-4f00-a78d-e33802c262f1-certs\") pod \"machine-config-server-ldxc8\" (UID: \"f3e58300-1dc1-4f00-a78d-e33802c262f1\") " pod="openshift-machine-config-operator/machine-config-server-ldxc8" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.380165 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tkfq\" (UniqueName: \"kubernetes.io/projected/7e3a84de-b4dd-4fc0-a279-9e2a9a616216-kube-api-access-6tkfq\") pod \"dns-default-cmb5q\" (UID: \"7e3a84de-b4dd-4fc0-a279-9e2a9a616216\") " pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.380343 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l7bk\" (UniqueName: \"kubernetes.io/projected/785b2c3d-cdc5-4c45-b4bb-3274cdb2a420-kube-api-access-2l7bk\") pod \"openshift-apiserver-operator-796bbdcf4f-nxv59\" (UID: \"785b2c3d-cdc5-4c45-b4bb-3274cdb2a420\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.380487 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4hhdc\" (UID: \"ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.381430 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpjq7\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-kube-api-access-xpjq7\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.385760 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a155afc3-d969-48f2-abc6-720c1f036e3c-cert\") pod \"ingress-canary-szpd6\" (UID: \"a155afc3-d969-48f2-abc6-720c1f036e3c\") " pod="openshift-ingress-canary/ingress-canary-szpd6" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.386517 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b8096afa-b0ad-4a6d-b301-108d746d6bbb-webhook-cert\") pod \"packageserver-d55dfcdfc-5z7tb\" (UID: \"b8096afa-b0ad-4a6d-b301-108d746d6bbb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.387980 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ztd4t"] Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.391997 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvbm2\" (UniqueName: \"kubernetes.io/projected/5473931a-2e46-45af-81f5-89fe0d299e4b-kube-api-access-tvbm2\") pod \"machine-config-controller-84d6567774-qrfsk\" (UID: \"5473931a-2e46-45af-81f5-89fe0d299e4b\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.400084 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.400336 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/437f0c85-bf20-4e24-9138-cce77c62fd2c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8k2xh\" (UID: \"437f0c85-bf20-4e24-9138-cce77c62fd2c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8k2xh" Jan 26 09:09:29 crc kubenswrapper[4872]: E0126 09:09:29.400915 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:29.900890477 +0000 UTC m=+103.209730268 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.402139 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7n6b\" (UniqueName: \"kubernetes.io/projected/1bd8496f-6499-4cf1-894a-37f68df536f4-kube-api-access-v7n6b\") pod \"migrator-59844c95c7-pplqc\" (UID: \"1bd8496f-6499-4cf1-894a-37f68df536f4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pplqc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.406847 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/437f0c85-bf20-4e24-9138-cce77c62fd2c-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-8k2xh\" (UID: \"437f0c85-bf20-4e24-9138-cce77c62fd2c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8k2xh" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.409574 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/835cb68e-1243-4522-bfd0-181378353e51-trusted-ca\") pod \"ingress-operator-5b745b69d9-nwkxd\" (UID: \"835cb68e-1243-4522-bfd0-181378353e51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.410084 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-bound-sa-token\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.412380 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/238152aa-010b-437e-8439-f8197970d5f9-srv-cert\") pod \"olm-operator-6b444d44fb-gfcvb\" (UID: \"238152aa-010b-437e-8439-f8197970d5f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.412759 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/031837e9-7dec-4179-9194-3b95092a8569-serving-cert\") pod \"service-ca-operator-777779d784-vvzw4\" (UID: \"031837e9-7dec-4179-9194-3b95092a8569\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.412777 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7975e724-e9f6-4a68-8779-f894ac687f74-serving-cert\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.413448 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/238152aa-010b-437e-8439-f8197970d5f9-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gfcvb\" (UID: \"238152aa-010b-437e-8439-f8197970d5f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.416311 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-metrics-certs\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.417016 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-stats-auth\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: W0126 09:09:29.425092 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d625f91_d4de_4687_bdbe_8ff985d7e406.slice/crio-346c36eba54b6a1cd4c7bbc2a46dbe0dfc15407ec6c13f11c6e4f54804ea3a1c WatchSource:0}: Error finding container 346c36eba54b6a1cd4c7bbc2a46dbe0dfc15407ec6c13f11c6e4f54804ea3a1c: Status 404 returned error can't find the container with id 346c36eba54b6a1cd4c7bbc2a46dbe0dfc15407ec6c13f11c6e4f54804ea3a1c Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.432840 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vpt9\" (UniqueName: \"kubernetes.io/projected/f2c67438-c0c7-48b8-a3df-21669ccb1602-kube-api-access-8vpt9\") pod \"collect-profiles-29490300-2v9qs\" (UID: \"f2c67438-c0c7-48b8-a3df-21669ccb1602\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.434365 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.436529 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/326f7b91-1e87-4fc0-9913-40f5f93d2cf2-default-certificate\") pod \"router-default-5444994796-9tdst\" (UID: \"326f7b91-1e87-4fc0-9913-40f5f93d2cf2\") " pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.458368 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.458940 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcr9k\" (UniqueName: \"kubernetes.io/projected/b8096afa-b0ad-4a6d-b301-108d746d6bbb-kube-api-access-rcr9k\") pod \"packageserver-d55dfcdfc-5z7tb\" (UID: \"b8096afa-b0ad-4a6d-b301-108d746d6bbb\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.465603 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv"] Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.472464 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5hh4\" (UniqueName: \"kubernetes.io/projected/031837e9-7dec-4179-9194-3b95092a8569-kube-api-access-v5hh4\") pod \"service-ca-operator-777779d784-vvzw4\" (UID: \"031837e9-7dec-4179-9194-3b95092a8569\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.492028 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-w5lkb"] Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.493730 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk"] Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.495970 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck78b\" (UniqueName: \"kubernetes.io/projected/ea814b88-1c4c-407b-8754-f57e19f8153f-kube-api-access-ck78b\") pod \"csi-hostpathplugin-lw526\" (UID: \"ea814b88-1c4c-407b-8754-f57e19f8153f\") " pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.501685 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:29 crc kubenswrapper[4872]: E0126 09:09:29.501894 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:30.0018553 +0000 UTC m=+103.310695101 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.502105 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: E0126 09:09:29.502691 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:30.002675533 +0000 UTC m=+103.311515334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.507430 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87dwr\" (UniqueName: \"kubernetes.io/projected/06e81723-7f8b-4dd6-99c6-d1ce1a966b7d-kube-api-access-87dwr\") pod \"kube-storage-version-migrator-operator-b67b599dd-zbn9p\" (UID: \"06e81723-7f8b-4dd6-99c6-d1ce1a966b7d\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.522616 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk"] Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.533205 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-4hhdc\" (UID: \"ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.543569 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg"] Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.553853 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l75vq\" (UniqueName: \"kubernetes.io/projected/835cb68e-1243-4522-bfd0-181378353e51-kube-api-access-l75vq\") pod \"ingress-operator-5b745b69d9-nwkxd\" (UID: \"835cb68e-1243-4522-bfd0-181378353e51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.569067 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-zr6z9" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.570836 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd54l\" (UniqueName: \"kubernetes.io/projected/55d2e904-b5e7-4bcf-8c39-1e84e51a2581-kube-api-access-xd54l\") pod \"package-server-manager-789f6589d5-dcs6q\" (UID: \"55d2e904-b5e7-4bcf-8c39-1e84e51a2581\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.606861 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.607451 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpthw\" (UniqueName: \"kubernetes.io/projected/a155afc3-d969-48f2-abc6-720c1f036e3c-kube-api-access-xpthw\") pod \"ingress-canary-szpd6\" (UID: \"a155afc3-d969-48f2-abc6-720c1f036e3c\") " pod="openshift-ingress-canary/ingress-canary-szpd6" Jan 26 09:09:29 crc kubenswrapper[4872]: E0126 09:09:29.607470 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:30.10744203 +0000 UTC m=+103.416281831 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.609234 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.616919 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/835cb68e-1243-4522-bfd0-181378353e51-bound-sa-token\") pod \"ingress-operator-5b745b69d9-nwkxd\" (UID: \"835cb68e-1243-4522-bfd0-181378353e51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.627237 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.631582 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.632023 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.638383 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8sfc\" (UniqueName: \"kubernetes.io/projected/f3e58300-1dc1-4f00-a78d-e33802c262f1-kube-api-access-x8sfc\") pod \"machine-config-server-ldxc8\" (UID: \"f3e58300-1dc1-4f00-a78d-e33802c262f1\") " pod="openshift-machine-config-operator/machine-config-server-ldxc8" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.638900 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.669471 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqr7m\" (UniqueName: \"kubernetes.io/projected/7975e724-e9f6-4a68-8779-f894ac687f74-kube-api-access-lqr7m\") pod \"authentication-operator-69f744f599-ntnp2\" (UID: \"7975e724-e9f6-4a68-8779-f894ac687f74\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.681356 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.691119 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.691180 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.698550 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pplqc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.707401 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjd9h\" (UniqueName: \"kubernetes.io/projected/238152aa-010b-437e-8439-f8197970d5f9-kube-api-access-xjd9h\") pod \"olm-operator-6b444d44fb-gfcvb\" (UID: \"238152aa-010b-437e-8439-f8197970d5f9\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.711704 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.714335 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf9fk\" (UniqueName: \"kubernetes.io/projected/437f0c85-bf20-4e24-9138-cce77c62fd2c-kube-api-access-tf9fk\") pod \"multus-admission-controller-857f4d67dd-8k2xh\" (UID: \"437f0c85-bf20-4e24-9138-cce77c62fd2c\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-8k2xh" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.715476 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.716487 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7bsh\" (UniqueName: \"kubernetes.io/projected/9b1a66ff-c739-4077-ab76-bf675e158a77-kube-api-access-l7bsh\") pod \"control-plane-machine-set-operator-78cbb6b69f-hvnl5\" (UID: \"9b1a66ff-c739-4077-ab76-bf675e158a77\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5" Jan 26 09:09:29 crc kubenswrapper[4872]: E0126 09:09:29.716564 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:30.216549717 +0000 UTC m=+103.525389518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.727576 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.729680 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.743619 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.751271 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.764050 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j98s\" (UniqueName: \"kubernetes.io/projected/a8658e2c-7305-4770-9b0d-1097def3df9f-kube-api-access-5j98s\") pod \"service-ca-9c57cc56f-n7pfp\" (UID: \"a8658e2c-7305-4770-9b0d-1097def3df9f\") " pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.787240 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-lw526" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.798482 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-ldxc8" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.808135 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-szpd6" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.820294 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:29 crc kubenswrapper[4872]: E0126 09:09:29.820481 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:30.320447221 +0000 UTC m=+103.629287022 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.820559 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:29 crc kubenswrapper[4872]: E0126 09:09:29.820997 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:30.320987576 +0000 UTC m=+103.629827377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.922483 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:29 crc kubenswrapper[4872]: E0126 09:09:29.924853 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:30.424833188 +0000 UTC m=+103.733672989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.925079 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.956112 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.963620 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" event={"ID":"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8","Type":"ContainerStarted","Data":"636065f9e0fb91ddd2e5023f8178193560c135de1cd75199d6ca76c7be759d58"} Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.964365 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.978109 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-8k2xh" Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.985114 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" event={"ID":"596ad6ba-b5ec-471c-96b6-1b24d525b163","Type":"ContainerStarted","Data":"e5819feef382d300f684035d9b58e6865d004da941fb9e9c5974b810413cea94"} Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.994773 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" event={"ID":"64013929-bfea-444c-8bf9-afbc7625a690","Type":"ContainerStarted","Data":"9742d423192f55007070859069b572d9121ba4a5a455f085e75b05b5edf8a6e4"} Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.994863 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" event={"ID":"64013929-bfea-444c-8bf9-afbc7625a690","Type":"ContainerStarted","Data":"19043407f34a98110f7c09a6eca92779855fa6bf74bdda68cff8e9f7eb8aa1b4"} Jan 26 09:09:29 crc kubenswrapper[4872]: I0126 09:09:29.998969 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" event={"ID":"bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1","Type":"ContainerStarted","Data":"66abbcaa5aa87c7f1ab10f34dd8681e4a8ba666078cd850a2a7badb00b17e143"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.016247 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" event={"ID":"d32d9b8b-2373-4c15-b5ce-f7f77c899cdd","Type":"ContainerStarted","Data":"2a185a5b4fa418c9dd114db6971c663b1a633f12373f01873cabf3d1efe532da"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.026587 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:30 crc kubenswrapper[4872]: E0126 09:09:30.026963 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:30.526952523 +0000 UTC m=+103.835792324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.028544 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.059770 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" event={"ID":"f8486391-70e4-4b2f-a093-ea9ab42c18bd","Type":"ContainerStarted","Data":"a7f9262791f406c11683e5803d7ab7d7c2a003fe1d081197f690d34f946781d7"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.066433 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" event={"ID":"586786e8-df0e-4424-8c84-bc2408b2dea5","Type":"ContainerStarted","Data":"7e0a68f7d10964b4569e82738c4f2707935e9aa7104d8f99440cc8b4e9c76e24"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.066476 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" event={"ID":"586786e8-df0e-4424-8c84-bc2408b2dea5","Type":"ContainerStarted","Data":"14d620988af2e378d67b4be6e2a3243c87686a3f218f37594b57e559fff2434b"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.067724 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" event={"ID":"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc","Type":"ContainerStarted","Data":"71adec89f961311a526882a35b558d02bf7374e9bdf0c458f96a0a0005a2a143"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.067744 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" event={"ID":"f7600bb7-8c57-41ac-afe1-f2c9eba5f7bc","Type":"ContainerStarted","Data":"2698d31092b07e47d2269d9a6bd6eff5a131d1f461f6d90d4d1466d0f89a1aec"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.105873 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ztd4t" event={"ID":"9d625f91-d4de-4687-bdbe-8ff985d7e406","Type":"ContainerStarted","Data":"346c36eba54b6a1cd4c7bbc2a46dbe0dfc15407ec6c13f11c6e4f54804ea3a1c"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.134417 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:30 crc kubenswrapper[4872]: E0126 09:09:30.135909 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:30.635888545 +0000 UTC m=+103.944728336 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.169217 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" event={"ID":"340d9cdb-6984-4ccb-aaa3-f7a0e7402021","Type":"ContainerStarted","Data":"04d07c6c6af6a1a8dab3cb3f42f6b306218b42ddfbf674947cd92dbdd1b6fca8"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.169979 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cmb5q"] Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.171759 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-w5lkb" event={"ID":"52b51ab7-5c1b-4cbb-a471-905a56f96a84","Type":"ContainerStarted","Data":"03597caefea79c3d51d96ee76fdf3a87040b5092789bff1ae79f1e307fa1f3ab"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.222242 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-82cml" event={"ID":"b24744f1-cd08-4ac4-b6c5-13613cfa88d0","Type":"ContainerStarted","Data":"f41225dc39cf9de3375dc8f8ec46c1c272371f9d59f3680c4ff5ad1363fe091d"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.238041 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.239279 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" event={"ID":"ebeadae0-aa3e-47d1-8068-1c339da55d91","Type":"ContainerStarted","Data":"643a21ff7402fe0cc5d8255da191b38d40a3974f7b23ccd3cbfd12c06041bd1d"} Jan 26 09:09:30 crc kubenswrapper[4872]: E0126 09:09:30.242084 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:30.742071851 +0000 UTC m=+104.050911652 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.260623 4872 generic.go:334] "Generic (PLEG): container finished" podID="5a111e46-9f4e-465f-821f-b58864174417" containerID="57dfe2fd2cf19b08ac8b9ffa514783e90228f5d7ee87ae9e2e53a9ddf0dae2f3" exitCode=0 Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.260744 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" event={"ID":"5a111e46-9f4e-465f-821f-b58864174417","Type":"ContainerDied","Data":"57dfe2fd2cf19b08ac8b9ffa514783e90228f5d7ee87ae9e2e53a9ddf0dae2f3"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.260783 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" event={"ID":"5a111e46-9f4e-465f-821f-b58864174417","Type":"ContainerStarted","Data":"2703cd6b133bb740a90156cf56c7a4eeca19801ac1a862af3115fc91e843566d"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.267170 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" event={"ID":"89ad9943-b0ac-4b88-aaf0-bdb332b22488","Type":"ContainerStarted","Data":"23f1688a97e3cd99170022305c5617ab26cb51aa20cc8f3a38c83d8f67802041"} Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.296963 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.307936 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bsfdw"] Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.383410 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:30 crc kubenswrapper[4872]: E0126 09:09:30.386141 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:30.886107457 +0000 UTC m=+104.194947438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.416703 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg"] Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.464464 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-zr6z9"] Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.499058 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:30 crc kubenswrapper[4872]: E0126 09:09:30.499518 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:30.999501761 +0000 UTC m=+104.308341572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.599641 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:30 crc kubenswrapper[4872]: E0126 09:09:30.599994 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:31.099979611 +0000 UTC m=+104.408819412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.600261 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" podStartSLOduration=84.600231097 podStartE2EDuration="1m24.600231097s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:30.598918432 +0000 UTC m=+103.907758243" watchObservedRunningTime="2026-01-26 09:09:30.600231097 +0000 UTC m=+103.909070898" Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.677623 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd"] Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.701828 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:30 crc kubenswrapper[4872]: E0126 09:09:30.702295 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:31.20228185 +0000 UTC m=+104.511121651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.803370 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:30 crc kubenswrapper[4872]: E0126 09:09:30.804050 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:31.304033625 +0000 UTC m=+104.612873426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.804152 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:30 crc kubenswrapper[4872]: E0126 09:09:30.804462 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:31.304455847 +0000 UTC m=+104.613295648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.908546 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:30 crc kubenswrapper[4872]: E0126 09:09:30.908837 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:31.408821654 +0000 UTC m=+104.717661455 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.923766 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59"] Jan 26 09:09:30 crc kubenswrapper[4872]: I0126 09:09:30.938971 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-th5xh" podStartSLOduration=84.938954571 podStartE2EDuration="1m24.938954571s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:30.937418679 +0000 UTC m=+104.246258480" watchObservedRunningTime="2026-01-26 09:09:30.938954571 +0000 UTC m=+104.247794372" Jan 26 09:09:31 crc kubenswrapper[4872]: W0126 09:09:31.000907 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc85d43e4_21f0_47fc_ae62_057de42e1935.slice/crio-a56f77adfff0c31b413bea8fd63990983022138b0905e7d8f97a82a35de51e7b WatchSource:0}: Error finding container a56f77adfff0c31b413bea8fd63990983022138b0905e7d8f97a82a35de51e7b: Status 404 returned error can't find the container with id a56f77adfff0c31b413bea8fd63990983022138b0905e7d8f97a82a35de51e7b Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.019136 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.019564 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:31.519550315 +0000 UTC m=+104.828390116 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.124724 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.124990 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:31.62497454 +0000 UTC m=+104.933814331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.152564 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p"] Jan 26 09:09:31 crc kubenswrapper[4872]: W0126 09:09:31.178767 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod785b2c3d_cdc5_4c45_b4bb_3274cdb2a420.slice/crio-6371b66f88eac044b7171c23cbcd3c1e8c1b091fd95d9d3a8c02175f5695dbc7 WatchSource:0}: Error finding container 6371b66f88eac044b7171c23cbcd3c1e8c1b091fd95d9d3a8c02175f5695dbc7: Status 404 returned error can't find the container with id 6371b66f88eac044b7171c23cbcd3c1e8c1b091fd95d9d3a8c02175f5695dbc7 Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.231894 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.232410 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:31.732393891 +0000 UTC m=+105.041233682 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:31 crc kubenswrapper[4872]: W0126 09:09:31.323311 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3e58300_1dc1_4f00_a78d_e33802c262f1.slice/crio-3257169aa4f6e551655b72e4d8b9930cd820985f93b27cdc85afb90d81a24582 WatchSource:0}: Error finding container 3257169aa4f6e551655b72e4d8b9930cd820985f93b27cdc85afb90d81a24582: Status 404 returned error can't find the container with id 3257169aa4f6e551655b72e4d8b9930cd820985f93b27cdc85afb90d81a24582 Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.334227 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.334462 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:31.834428113 +0000 UTC m=+105.143267914 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.335128 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.340971 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:31.840942521 +0000 UTC m=+105.149782332 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.415540 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" event={"ID":"340d9cdb-6984-4ccb-aaa3-f7a0e7402021","Type":"ContainerStarted","Data":"3cabd4aef887e27d90a82ee7bdbfba628e04059058dfa0a1169fc8784b223a73"} Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.436534 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb"] Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.436956 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.437541 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:31.937517984 +0000 UTC m=+105.246357785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.463198 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ztd4t" event={"ID":"9d625f91-d4de-4687-bdbe-8ff985d7e406","Type":"ContainerStarted","Data":"cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723"} Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.480661 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" event={"ID":"835cb68e-1243-4522-bfd0-181378353e51","Type":"ContainerStarted","Data":"8c325ddb34b5ace85cfeface442b386f6adfea1cde60672fc8ad81ad10888584"} Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.493589 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" event={"ID":"785b2c3d-cdc5-4c45-b4bb-3274cdb2a420","Type":"ContainerStarted","Data":"6371b66f88eac044b7171c23cbcd3c1e8c1b091fd95d9d3a8c02175f5695dbc7"} Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.504626 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q"] Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.508297 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk"] Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.519229 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" event={"ID":"07b7e8f0-a421-4fbc-9c0d-aa773210edcd","Type":"ContainerStarted","Data":"b1d1246dd553dbdaccfa0ceed7f1dcbacd99d48b9c2dd373fc9243873f273591"} Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.540218 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9tdst" event={"ID":"326f7b91-1e87-4fc0-9913-40f5f93d2cf2","Type":"ContainerStarted","Data":"bc8a682f25e8474feaa38839570088a37746a6d45adb45e4cc3f4c1387bf9f0b"} Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.540957 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.543317 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:32.04330385 +0000 UTC m=+105.352143651 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.546665 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-x9zlw" podStartSLOduration=85.546642031 podStartE2EDuration="1m25.546642031s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:31.515544567 +0000 UTC m=+104.824384368" watchObservedRunningTime="2026-01-26 09:09:31.546642031 +0000 UTC m=+104.855481832" Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.548266 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5"] Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.568446 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cmb5q" event={"ID":"7e3a84de-b4dd-4fc0-a279-9e2a9a616216","Type":"ContainerStarted","Data":"1990ad0d423f38009073cdc0f2ebc081b81927f00db2b3490ed97c6f6ca71a52"} Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.648399 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-82cml" event={"ID":"b24744f1-cd08-4ac4-b6c5-13613cfa88d0","Type":"ContainerStarted","Data":"a1f6bdf39210155669bd9c6f4a396f3ac5bf6c0d3e48e3765c4c0e1d0f00fa20"} Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.650843 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.652264 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:32.152241442 +0000 UTC m=+105.461081243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.668835 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" podStartSLOduration=85.668782006 podStartE2EDuration="1m25.668782006s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:31.662196765 +0000 UTC m=+104.971036566" watchObservedRunningTime="2026-01-26 09:09:31.668782006 +0000 UTC m=+104.977621807" Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.711156 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" event={"ID":"596ad6ba-b5ec-471c-96b6-1b24d525b163","Type":"ContainerStarted","Data":"a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf"} Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.712145 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.713217 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-82cml" podStartSLOduration=85.713204686 podStartE2EDuration="1m25.713204686s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:31.711630093 +0000 UTC m=+105.020469894" watchObservedRunningTime="2026-01-26 09:09:31.713204686 +0000 UTC m=+105.022044487" Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.722330 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-zr6z9" event={"ID":"c85d43e4-21f0-47fc-ae62-057de42e1935","Type":"ContainerStarted","Data":"a56f77adfff0c31b413bea8fd63990983022138b0905e7d8f97a82a35de51e7b"} Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.754169 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.754644 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:32.254626494 +0000 UTC m=+105.563466295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.812186 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" podStartSLOduration=85.812160804 podStartE2EDuration="1m25.812160804s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:31.752605828 +0000 UTC m=+105.061445629" watchObservedRunningTime="2026-01-26 09:09:31.812160804 +0000 UTC m=+105.121000605" Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.860399 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.860648 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:32.360612315 +0000 UTC m=+105.669452116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.861340 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5dgh7" podStartSLOduration=85.861325574 podStartE2EDuration="1m25.861325574s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:31.814056356 +0000 UTC m=+105.122896157" watchObservedRunningTime="2026-01-26 09:09:31.861325574 +0000 UTC m=+105.170165375" Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.861636 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.862747 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-zw7z6" podStartSLOduration=85.862741063 podStartE2EDuration="1m25.862741063s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:31.860743079 +0000 UTC m=+105.169582880" watchObservedRunningTime="2026-01-26 09:09:31.862741063 +0000 UTC m=+105.171580864" Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.863750 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:32.363740991 +0000 UTC m=+105.672580782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.966684 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.967532 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:32.46749776 +0000 UTC m=+105.776337561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:31 crc kubenswrapper[4872]: I0126 09:09:31.980227 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:31 crc kubenswrapper[4872]: E0126 09:09:31.980754 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:32.480740464 +0000 UTC m=+105.789580265 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.082987 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:32 crc kubenswrapper[4872]: E0126 09:09:32.083114 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:32.583086206 +0000 UTC m=+105.891926017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.085599 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" podStartSLOduration=86.085568213 podStartE2EDuration="1m26.085568213s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:32.069650006 +0000 UTC m=+105.378489807" watchObservedRunningTime="2026-01-26 09:09:32.085568213 +0000 UTC m=+105.394408014" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.091611 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:32 crc kubenswrapper[4872]: E0126 09:09:32.093370 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:32.593347847 +0000 UTC m=+105.902187648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.106713 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-hx6pl" podStartSLOduration=86.106685623 podStartE2EDuration="1m26.106685623s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:32.105528412 +0000 UTC m=+105.414368223" watchObservedRunningTime="2026-01-26 09:09:32.106685623 +0000 UTC m=+105.415525414" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.175651 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-ztd4t" podStartSLOduration=86.175629847 podStartE2EDuration="1m26.175629847s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:32.161432186 +0000 UTC m=+105.470271997" watchObservedRunningTime="2026-01-26 09:09:32.175629847 +0000 UTC m=+105.484469648" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.194069 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:32 crc kubenswrapper[4872]: E0126 09:09:32.194369 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:32.694351891 +0000 UTC m=+106.003191692 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.295972 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:32 crc kubenswrapper[4872]: E0126 09:09:32.296415 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:32.796402354 +0000 UTC m=+106.105242155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.401053 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:32 crc kubenswrapper[4872]: E0126 09:09:32.401589 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:32.901566342 +0000 UTC m=+106.210406143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.452929 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb"] Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.503541 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:32 crc kubenswrapper[4872]: E0126 09:09:32.504228 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:33.004205831 +0000 UTC m=+106.313045632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.569732 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-szpd6"] Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.575477 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pplqc"] Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.604916 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:32 crc kubenswrapper[4872]: E0126 09:09:32.605159 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:33.105115452 +0000 UTC m=+106.413955263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.605434 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:32 crc kubenswrapper[4872]: E0126 09:09:32.606017 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:33.105996757 +0000 UTC m=+106.414836558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:32 crc kubenswrapper[4872]: W0126 09:09:32.665402 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda155afc3_d969_48f2_abc6_720c1f036e3c.slice/crio-2ea4b19658120d5466dd7170c0ef9299552bae0dcc7cdfa99219bd0d4d885b9e WatchSource:0}: Error finding container 2ea4b19658120d5466dd7170c0ef9299552bae0dcc7cdfa99219bd0d4d885b9e: Status 404 returned error can't find the container with id 2ea4b19658120d5466dd7170c0ef9299552bae0dcc7cdfa99219bd0d4d885b9e Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.696351 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc"] Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.707138 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:32 crc kubenswrapper[4872]: E0126 09:09:32.707482 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:33.207465203 +0000 UTC m=+106.516305004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.714030 4872 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-mk4bd container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.26:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.714068 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" podUID="596ad6ba-b5ec-471c-96b6-1b24d525b163" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.26:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.716601 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4"] Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.750086 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" event={"ID":"89ad9943-b0ac-4b88-aaf0-bdb332b22488","Type":"ContainerStarted","Data":"2e7e351cb176261590edfb718cf78d696e46f8391dd3057321a76bdf37ce043d"} Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.750715 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lw526"] Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.760638 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-n7pfp"] Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.766949 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" event={"ID":"55d2e904-b5e7-4bcf-8c39-1e84e51a2581","Type":"ContainerStarted","Data":"2ad6aa86ff4f8123b610a32f4d2cd6d6e4b4c3335c98cc8894d0fe8d2c36c01c"} Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.784428 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pplqc" event={"ID":"1bd8496f-6499-4cf1-894a-37f68df536f4","Type":"ContainerStarted","Data":"f197522285cd80da57b5d5e5069220825694ad49a494099dd5d2b6c43c8810ac"} Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.786032 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.786250 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.795222 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-g42sk" podStartSLOduration=86.795196873 podStartE2EDuration="1m26.795196873s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:32.793715993 +0000 UTC m=+106.102555794" watchObservedRunningTime="2026-01-26 09:09:32.795196873 +0000 UTC m=+106.104036674" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.803421 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-w5lkb" event={"ID":"52b51ab7-5c1b-4cbb-a471-905a56f96a84","Type":"ContainerStarted","Data":"56d9b5c127e889f1f982906ec936f12c1b9f945606996651e0951fdb4b316a6b"} Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.808365 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.809115 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:32 crc kubenswrapper[4872]: E0126 09:09:32.809748 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:33.309733712 +0000 UTC m=+106.618573513 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.814941 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.815262 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.828890 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-8k2xh"] Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.842563 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-ntnp2"] Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.854076 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-szpd6" event={"ID":"a155afc3-d969-48f2-abc6-720c1f036e3c","Type":"ContainerStarted","Data":"2ea4b19658120d5466dd7170c0ef9299552bae0dcc7cdfa99219bd0d4d885b9e"} Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.866911 4872 patch_prober.go:28] interesting pod/apiserver-76f77b778f-82cml container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Jan 26 09:09:32 crc kubenswrapper[4872]: [+]log ok Jan 26 09:09:32 crc kubenswrapper[4872]: [+]etcd ok Jan 26 09:09:32 crc kubenswrapper[4872]: [+]poststarthook/start-apiserver-admission-initializer ok Jan 26 09:09:32 crc kubenswrapper[4872]: [+]poststarthook/generic-apiserver-start-informers ok Jan 26 09:09:32 crc kubenswrapper[4872]: [+]poststarthook/max-in-flight-filter ok Jan 26 09:09:32 crc kubenswrapper[4872]: [+]poststarthook/storage-object-count-tracker-hook ok Jan 26 09:09:32 crc kubenswrapper[4872]: [+]poststarthook/image.openshift.io-apiserver-caches ok Jan 26 09:09:32 crc kubenswrapper[4872]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Jan 26 09:09:32 crc kubenswrapper[4872]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Jan 26 09:09:32 crc kubenswrapper[4872]: [+]poststarthook/project.openshift.io-projectcache ok Jan 26 09:09:32 crc kubenswrapper[4872]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Jan 26 09:09:32 crc kubenswrapper[4872]: [+]poststarthook/openshift.io-startinformers ok Jan 26 09:09:32 crc kubenswrapper[4872]: [+]poststarthook/openshift.io-restmapperupdater ok Jan 26 09:09:32 crc kubenswrapper[4872]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Jan 26 09:09:32 crc kubenswrapper[4872]: livez check failed Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.866975 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-82cml" podUID="b24744f1-cd08-4ac4-b6c5-13613cfa88d0" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.880433 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" event={"ID":"238152aa-010b-437e-8439-f8197970d5f9","Type":"ContainerStarted","Data":"c5582bcbea94abe8de160f2391ad10da937dffd61f43cb13f1153d439bde0d17"} Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.893551 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-zr6z9" event={"ID":"c85d43e4-21f0-47fc-ae62-057de42e1935","Type":"ContainerStarted","Data":"72ffac10a142cc1df6b418c2135a8e93c97ecaf54071819f7130569dadb0913f"} Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.894118 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs"] Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.894369 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-zr6z9" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.907392 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-zr6z9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.907479 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zr6z9" podUID="c85d43e4-21f0-47fc-ae62-057de42e1935" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.909726 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:32 crc kubenswrapper[4872]: E0126 09:09:32.911460 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:33.411418405 +0000 UTC m=+106.720258396 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.911925 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" event={"ID":"835cb68e-1243-4522-bfd0-181378353e51","Type":"ContainerStarted","Data":"283cc027452ec28ba88f1747eb65a7426b7685b79130fdab6a61ebb47c02ac35"} Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.921747 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-zr6z9" podStartSLOduration=86.921714538 podStartE2EDuration="1m26.921714538s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:32.921258626 +0000 UTC m=+106.230098427" watchObservedRunningTime="2026-01-26 09:09:32.921714538 +0000 UTC m=+106.230554339" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.926389 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" event={"ID":"bdd03aae-c256-4e3f-b7e9-a350c2fdf4c1","Type":"ContainerStarted","Data":"152ca7ea5b9bfc3c38cab03fdfadbffb3680f6e0a8c2412251b15110f3684e75"} Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.931451 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ldxc8" event={"ID":"f3e58300-1dc1-4f00-a78d-e33802c262f1","Type":"ContainerStarted","Data":"3257169aa4f6e551655b72e4d8b9930cd820985f93b27cdc85afb90d81a24582"} Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.933904 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" event={"ID":"d32d9b8b-2373-4c15-b5ce-f7f77c899cdd","Type":"ContainerStarted","Data":"824a2af076130f4a311ae994988abe11b067d47af1335043abdc3bbcac89d1a8"} Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.938891 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.974671 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" Jan 26 09:09:32 crc kubenswrapper[4872]: I0126 09:09:32.993965 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" event={"ID":"586786e8-df0e-4424-8c84-bc2408b2dea5","Type":"ContainerStarted","Data":"1ce65993979df16f1d95b82bb6327fda6016744df0524d7b9bae040d85d8938f"} Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.007783 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" event={"ID":"5473931a-2e46-45af-81f5-89fe0d299e4b","Type":"ContainerStarted","Data":"b76bb9aa8a2cf3c571eae2a0a0827c18aa7225b0c157727cc35b88855f391ea8"} Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.012380 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:33 crc kubenswrapper[4872]: E0126 09:09:33.014686 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:33.514674701 +0000 UTC m=+106.823514502 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.020149 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" event={"ID":"b8096afa-b0ad-4a6d-b301-108d746d6bbb","Type":"ContainerStarted","Data":"27d03c22d6de5042cf7d3e3eaf9aea07002cc7478d84c4094f1504c331910583"} Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.038406 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" event={"ID":"5a111e46-9f4e-465f-821f-b58864174417","Type":"ContainerStarted","Data":"2bd65f57c232776e8e9419bc53d67f88584e41e17c9f055bec15085624de592c"} Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.039026 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.047602 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg" event={"ID":"49b97b45-be77-455a-9a69-f20629af5940","Type":"ContainerStarted","Data":"ed6a1627e009f5242cbe5be69e413aa1ac6f791e7ca3aa624bdeb037d541eaac"} Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.053200 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-r2xzk" podStartSLOduration=87.053184739 podStartE2EDuration="1m27.053184739s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:32.958852788 +0000 UTC m=+106.267692589" watchObservedRunningTime="2026-01-26 09:09:33.053184739 +0000 UTC m=+106.362024540" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.053597 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6bfrv" podStartSLOduration=87.053591921 podStartE2EDuration="1m27.053591921s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:33.053407286 +0000 UTC m=+106.362247087" watchObservedRunningTime="2026-01-26 09:09:33.053591921 +0000 UTC m=+106.362431722" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.073522 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" event={"ID":"ebeadae0-aa3e-47d1-8068-1c339da55d91","Type":"ContainerStarted","Data":"a2fed046a8be715fc3f0afa4e08bf35aa815d323ac17fb6697fbed9c8fdd3912"} Jan 26 09:09:33 crc kubenswrapper[4872]: W0126 09:09:33.080121 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod437f0c85_bf20_4e24_9138_cce77c62fd2c.slice/crio-6c51bcf1a42f8a1e972830a84733b1613a8660c90056421ff212ca07d8ecb84b WatchSource:0}: Error finding container 6c51bcf1a42f8a1e972830a84733b1613a8660c90056421ff212ca07d8ecb84b: Status 404 returned error can't find the container with id 6c51bcf1a42f8a1e972830a84733b1613a8660c90056421ff212ca07d8ecb84b Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.081588 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5" event={"ID":"9b1a66ff-c739-4077-ab76-bf675e158a77","Type":"ContainerStarted","Data":"620fed0b2d53daf2d5cc52d1e841f8b9b0729d51efda4f808ef93e4854b00e49"} Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.108562 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" podStartSLOduration=87.1085484 podStartE2EDuration="1m27.1085484s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:33.105366383 +0000 UTC m=+106.414206184" watchObservedRunningTime="2026-01-26 09:09:33.1085484 +0000 UTC m=+106.417388191" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.115093 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" event={"ID":"06e81723-7f8b-4dd6-99c6-d1ce1a966b7d","Type":"ContainerStarted","Data":"fc207ee5086ecfa508f0ec6b36b816f39f8ab18d1c1122b39ee9ee5d63b745eb"} Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.130336 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:33 crc kubenswrapper[4872]: E0126 09:09:33.131859 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:33.63184562 +0000 UTC m=+106.940685411 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.154336 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.208049 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-qglzk" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.216809 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-c4vmf" podStartSLOduration=87.216761972 podStartE2EDuration="1m27.216761972s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:33.212300549 +0000 UTC m=+106.521140370" watchObservedRunningTime="2026-01-26 09:09:33.216761972 +0000 UTC m=+106.525601773" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.235667 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:33 crc kubenswrapper[4872]: E0126 09:09:33.236030 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:33.736017841 +0000 UTC m=+107.044857632 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.336854 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:33 crc kubenswrapper[4872]: E0126 09:09:33.337508 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:33.837482307 +0000 UTC m=+107.146322108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.438099 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:33 crc kubenswrapper[4872]: E0126 09:09:33.438514 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:33.938494262 +0000 UTC m=+107.247334063 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.539671 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:33 crc kubenswrapper[4872]: E0126 09:09:33.540424 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:34.04037735 +0000 UTC m=+107.349217151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.542098 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5xspp"] Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.543206 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.546436 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.563631 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5xspp"] Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.647004 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.647081 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-catalog-content\") pod \"community-operators-5xspp\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.647121 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-utilities\") pod \"community-operators-5xspp\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.647161 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br9rq\" (UniqueName: \"kubernetes.io/projected/7b6eb987-454b-4c50-bb39-ede2e1006808-kube-api-access-br9rq\") pod \"community-operators-5xspp\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:09:33 crc kubenswrapper[4872]: E0126 09:09:33.647599 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:34.147587685 +0000 UTC m=+107.456427486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.748683 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.748999 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-catalog-content\") pod \"community-operators-5xspp\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.749029 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-utilities\") pod \"community-operators-5xspp\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.749070 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br9rq\" (UniqueName: \"kubernetes.io/projected/7b6eb987-454b-4c50-bb39-ede2e1006808-kube-api-access-br9rq\") pod \"community-operators-5xspp\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:09:33 crc kubenswrapper[4872]: E0126 09:09:33.749422 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:34.249408072 +0000 UTC m=+107.558247873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.749731 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-catalog-content\") pod \"community-operators-5xspp\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.749955 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-utilities\") pod \"community-operators-5xspp\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.750907 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p76kx"] Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.752100 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.757061 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.802155 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br9rq\" (UniqueName: \"kubernetes.io/projected/7b6eb987-454b-4c50-bb39-ede2e1006808-kube-api-access-br9rq\") pod \"community-operators-5xspp\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.841724 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p76kx"] Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.855078 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c963ddba-b620-46b4-83af-dfbaa96bbd03-utilities\") pod \"certified-operators-p76kx\" (UID: \"c963ddba-b620-46b4-83af-dfbaa96bbd03\") " pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.855651 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.855754 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c963ddba-b620-46b4-83af-dfbaa96bbd03-catalog-content\") pod \"certified-operators-p76kx\" (UID: \"c963ddba-b620-46b4-83af-dfbaa96bbd03\") " pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.855868 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrrwv\" (UniqueName: \"kubernetes.io/projected/c963ddba-b620-46b4-83af-dfbaa96bbd03-kube-api-access-vrrwv\") pod \"certified-operators-p76kx\" (UID: \"c963ddba-b620-46b4-83af-dfbaa96bbd03\") " pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:09:33 crc kubenswrapper[4872]: E0126 09:09:33.856306 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:34.356293117 +0000 UTC m=+107.665132918 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.931986 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rng9p"] Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.955299 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.958625 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.959117 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c963ddba-b620-46b4-83af-dfbaa96bbd03-utilities\") pod \"certified-operators-p76kx\" (UID: \"c963ddba-b620-46b4-83af-dfbaa96bbd03\") " pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.959216 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c963ddba-b620-46b4-83af-dfbaa96bbd03-catalog-content\") pod \"certified-operators-p76kx\" (UID: \"c963ddba-b620-46b4-83af-dfbaa96bbd03\") " pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.959272 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrrwv\" (UniqueName: \"kubernetes.io/projected/c963ddba-b620-46b4-83af-dfbaa96bbd03-kube-api-access-vrrwv\") pod \"certified-operators-p76kx\" (UID: \"c963ddba-b620-46b4-83af-dfbaa96bbd03\") " pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:09:33 crc kubenswrapper[4872]: E0126 09:09:33.959844 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:34.459818231 +0000 UTC m=+107.768658032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.960335 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c963ddba-b620-46b4-83af-dfbaa96bbd03-utilities\") pod \"certified-operators-p76kx\" (UID: \"c963ddba-b620-46b4-83af-dfbaa96bbd03\") " pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.960598 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c963ddba-b620-46b4-83af-dfbaa96bbd03-catalog-content\") pod \"certified-operators-p76kx\" (UID: \"c963ddba-b620-46b4-83af-dfbaa96bbd03\") " pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:09:33 crc kubenswrapper[4872]: I0126 09:09:33.978093 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rng9p"] Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.003681 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.013256 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrrwv\" (UniqueName: \"kubernetes.io/projected/c963ddba-b620-46b4-83af-dfbaa96bbd03-kube-api-access-vrrwv\") pod \"certified-operators-p76kx\" (UID: \"c963ddba-b620-46b4-83af-dfbaa96bbd03\") " pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.062406 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bca70073-3d97-452c-818c-70c8021fe37d-utilities\") pod \"community-operators-rng9p\" (UID: \"bca70073-3d97-452c-818c-70c8021fe37d\") " pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.062958 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.063004 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtx4m\" (UniqueName: \"kubernetes.io/projected/bca70073-3d97-452c-818c-70c8021fe37d-kube-api-access-mtx4m\") pod \"community-operators-rng9p\" (UID: \"bca70073-3d97-452c-818c-70c8021fe37d\") " pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.063030 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bca70073-3d97-452c-818c-70c8021fe37d-catalog-content\") pod \"community-operators-rng9p\" (UID: \"bca70073-3d97-452c-818c-70c8021fe37d\") " pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:09:34 crc kubenswrapper[4872]: E0126 09:09:34.063568 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:34.56355006 +0000 UTC m=+107.872389861 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.105120 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.136146 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p5rmt"] Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.137317 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.150418 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p5rmt"] Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.164479 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.164864 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtx4m\" (UniqueName: \"kubernetes.io/projected/bca70073-3d97-452c-818c-70c8021fe37d-kube-api-access-mtx4m\") pod \"community-operators-rng9p\" (UID: \"bca70073-3d97-452c-818c-70c8021fe37d\") " pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.164901 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bca70073-3d97-452c-818c-70c8021fe37d-catalog-content\") pod \"community-operators-rng9p\" (UID: \"bca70073-3d97-452c-818c-70c8021fe37d\") " pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.164949 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnswd\" (UniqueName: \"kubernetes.io/projected/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-kube-api-access-fnswd\") pod \"certified-operators-p5rmt\" (UID: \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\") " pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.165025 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bca70073-3d97-452c-818c-70c8021fe37d-utilities\") pod \"community-operators-rng9p\" (UID: \"bca70073-3d97-452c-818c-70c8021fe37d\") " pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.165111 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-catalog-content\") pod \"certified-operators-p5rmt\" (UID: \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\") " pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.165139 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-utilities\") pod \"certified-operators-p5rmt\" (UID: \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\") " pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.165548 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bca70073-3d97-452c-818c-70c8021fe37d-catalog-content\") pod \"community-operators-rng9p\" (UID: \"bca70073-3d97-452c-818c-70c8021fe37d\") " pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.165589 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bca70073-3d97-452c-818c-70c8021fe37d-utilities\") pod \"community-operators-rng9p\" (UID: \"bca70073-3d97-452c-818c-70c8021fe37d\") " pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:09:34 crc kubenswrapper[4872]: E0126 09:09:34.165643 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:34.665622363 +0000 UTC m=+107.974462164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.230093 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" event={"ID":"7975e724-e9f6-4a68-8779-f894ac687f74","Type":"ContainerStarted","Data":"04e068803b494c8aad6018535da4ddd2541f79ce2ebff81d349591685f7f026f"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.239517 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtx4m\" (UniqueName: \"kubernetes.io/projected/bca70073-3d97-452c-818c-70c8021fe37d-kube-api-access-mtx4m\") pod \"community-operators-rng9p\" (UID: \"bca70073-3d97-452c-818c-70c8021fe37d\") " pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.265156 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-w5lkb" event={"ID":"52b51ab7-5c1b-4cbb-a471-905a56f96a84","Type":"ContainerStarted","Data":"44fe451c7583f6cd168f4b86c89ac6b650d71a529f547fc38d1de3f6ad03a51d"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.266725 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.266767 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-catalog-content\") pod \"certified-operators-p5rmt\" (UID: \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\") " pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.266787 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-utilities\") pod \"certified-operators-p5rmt\" (UID: \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\") " pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.266871 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnswd\" (UniqueName: \"kubernetes.io/projected/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-kube-api-access-fnswd\") pod \"certified-operators-p5rmt\" (UID: \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\") " pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:09:34 crc kubenswrapper[4872]: E0126 09:09:34.268307 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:34.768292143 +0000 UTC m=+108.077131944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.273054 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-catalog-content\") pod \"certified-operators-p5rmt\" (UID: \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\") " pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.273334 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lw526" event={"ID":"ea814b88-1c4c-407b-8754-f57e19f8153f","Type":"ContainerStarted","Data":"d5ff5d1c2b5251d622e4dd2598615da9daad6cc26836723f96bf6b4531a794d2"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.273341 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-utilities\") pod \"certified-operators-p5rmt\" (UID: \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\") " pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.290987 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg" event={"ID":"49b97b45-be77-455a-9a69-f20629af5940","Type":"ContainerStarted","Data":"9112d9c568e76d902eec70fe5caa3c84c0d64c7150e297613b0c5b163ef6efbe"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.312116 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-w5lkb" podStartSLOduration=88.312095276 podStartE2EDuration="1m28.312095276s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.29877151 +0000 UTC m=+107.607611311" watchObservedRunningTime="2026-01-26 09:09:34.312095276 +0000 UTC m=+107.620935077" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.322390 4872 csr.go:261] certificate signing request csr-8d9jx is approved, waiting to be issued Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.323564 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" event={"ID":"5473931a-2e46-45af-81f5-89fe0d299e4b","Type":"ContainerStarted","Data":"2f260eea7182f72e883e015afbfea4447102749bf4b29049dfb7c3b30d05bf61"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.323643 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" event={"ID":"5473931a-2e46-45af-81f5-89fe0d299e4b","Type":"ContainerStarted","Data":"0abcf9116db47b4fd3f18f78e33be7a585e978a5503861fe86cafe17779f3ad6"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.332063 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnswd\" (UniqueName: \"kubernetes.io/projected/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-kube-api-access-fnswd\") pod \"certified-operators-p5rmt\" (UID: \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\") " pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.349010 4872 csr.go:257] certificate signing request csr-8d9jx is issued Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.358523 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-qrfsk" podStartSLOduration=88.358492271 podStartE2EDuration="1m28.358492271s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.358220452 +0000 UTC m=+107.667060253" watchObservedRunningTime="2026-01-26 09:09:34.358492271 +0000 UTC m=+107.667332072" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.367614 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:34 crc kubenswrapper[4872]: E0126 09:09:34.372086 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:34.872056133 +0000 UTC m=+108.180895934 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.375194 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.388171 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" event={"ID":"785b2c3d-cdc5-4c45-b4bb-3274cdb2a420","Type":"ContainerStarted","Data":"d57dddd7649c4c4b29e19b30cce1fe07f2ccbd5f81e88c0dfe0f0ee3ef604cea"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.417189 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-ldxc8" event={"ID":"f3e58300-1dc1-4f00-a78d-e33802c262f1","Type":"ContainerStarted","Data":"bb307c0d2384ab49c8a44dda6377022ff4f09af19219f270503c55f3d664415d"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.427090 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cmb5q" event={"ID":"7e3a84de-b4dd-4fc0-a279-9e2a9a616216","Type":"ContainerStarted","Data":"8ddde279ffd2c107c749e07bc3d6620fe2ac3c11ff9c625c3b1e0b855fc902f1"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.444474 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" event={"ID":"b8096afa-b0ad-4a6d-b301-108d746d6bbb","Type":"ContainerStarted","Data":"bbf108543862008c6dae435fa16fb474b57d9dfc5c0ba3ad9107a0f2a9a7552c"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.445391 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.446598 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-nxv59" podStartSLOduration=88.44657148 podStartE2EDuration="1m28.44657148s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.438246181 +0000 UTC m=+107.747085992" watchObservedRunningTime="2026-01-26 09:09:34.44657148 +0000 UTC m=+107.755411271" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.467773 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" event={"ID":"ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6","Type":"ContainerStarted","Data":"ccd1ac94f7588f2b937ca1dcbdd428504ed1d9f6a42a29d0950aeac2282f3a38"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.469872 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:34 crc kubenswrapper[4872]: E0126 09:09:34.471783 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:34.971768662 +0000 UTC m=+108.280608463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.476558 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-ldxc8" podStartSLOduration=8.476539583 podStartE2EDuration="8.476539583s" podCreationTimestamp="2026-01-26 09:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.475757551 +0000 UTC m=+107.784597372" watchObservedRunningTime="2026-01-26 09:09:34.476539583 +0000 UTC m=+107.785379384" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.506714 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" event={"ID":"f2c67438-c0c7-48b8-a3df-21669ccb1602","Type":"ContainerStarted","Data":"76664482753c3c2630369c1de9f8fe161bea6d4557a168d0d81b98f6c2e95d4c"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.507237 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" event={"ID":"f2c67438-c0c7-48b8-a3df-21669ccb1602","Type":"ContainerStarted","Data":"77d3c90487255fd31a97c328df74ad31f693023f2b3e0048e4861275ef2b0313"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.519397 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" podStartSLOduration=88.519372759 podStartE2EDuration="1m28.519372759s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.516463729 +0000 UTC m=+107.825303530" watchObservedRunningTime="2026-01-26 09:09:34.519372759 +0000 UTC m=+107.828212560" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.529095 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.563017 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-szpd6" event={"ID":"a155afc3-d969-48f2-abc6-720c1f036e3c","Type":"ContainerStarted","Data":"ace1b687d4ea6305e21202123c64643337ad2dec70214d8647a07e2d72c9bf7a"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.575021 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:34 crc kubenswrapper[4872]: E0126 09:09:34.576154 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:35.076129538 +0000 UTC m=+108.384969339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.589033 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pplqc" event={"ID":"1bd8496f-6499-4cf1-894a-37f68df536f4","Type":"ContainerStarted","Data":"561e9b85f9fb96ea5cee0767d26f78bdb83e65b55c210e56fabc9c8791682e3c"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.608559 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8k2xh" event={"ID":"437f0c85-bf20-4e24-9138-cce77c62fd2c","Type":"ContainerStarted","Data":"6c51bcf1a42f8a1e972830a84733b1613a8660c90056421ff212ca07d8ecb84b"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.614598 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" event={"ID":"a8658e2c-7305-4770-9b0d-1097def3df9f","Type":"ContainerStarted","Data":"93ac8c6dc2a6c1ecab4d4ccb9e556b39df91c31db7da8f09929adb4701180426"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.614676 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" event={"ID":"a8658e2c-7305-4770-9b0d-1097def3df9f","Type":"ContainerStarted","Data":"e802776226254b92797610e009e2333d21e0b38c2c75d9dd49e34ad23b45cee8"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.617934 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" podStartSLOduration=88.617909255 podStartE2EDuration="1m28.617909255s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.54818416 +0000 UTC m=+107.857023961" watchObservedRunningTime="2026-01-26 09:09:34.617909255 +0000 UTC m=+107.926749056" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.665525 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" event={"ID":"031837e9-7dec-4179-9194-3b95092a8569","Type":"ContainerStarted","Data":"8a3a234a76d9fd374d49907ac81a4c935683a4ca4b93fba7a6525cbb1883165e"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.665594 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" event={"ID":"031837e9-7dec-4179-9194-3b95092a8569","Type":"ContainerStarted","Data":"df9ae84f05719bc22de7a5f340182957ce33b7ea95975837816b4d9974e9f069"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.685275 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" event={"ID":"238152aa-010b-437e-8439-f8197970d5f9","Type":"ContainerStarted","Data":"fa15e40941c4ca7d9854111c98f8d1ee4aca3b7296030cf4d1b1e5cd8bbd4209"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.686340 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.691233 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:34 crc kubenswrapper[4872]: E0126 09:09:34.693562 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:35.193540213 +0000 UTC m=+108.502380014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.711080 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-szpd6" podStartSLOduration=8.711046673 podStartE2EDuration="8.711046673s" podCreationTimestamp="2026-01-26 09:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.620060514 +0000 UTC m=+107.928900315" watchObservedRunningTime="2026-01-26 09:09:34.711046673 +0000 UTC m=+108.019886474" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.733346 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pplqc" podStartSLOduration=88.733316595 podStartE2EDuration="1m28.733316595s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.695388623 +0000 UTC m=+108.004228434" watchObservedRunningTime="2026-01-26 09:09:34.733316595 +0000 UTC m=+108.042156396" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.736307 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.743276 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5" event={"ID":"9b1a66ff-c739-4077-ab76-bf675e158a77","Type":"ContainerStarted","Data":"5f7a0846b27e794364478e63360fd1b75cd1cb6acd8dacfdce859c8490379517"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.754386 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-n7pfp" podStartSLOduration=88.754354222 podStartE2EDuration="1m28.754354222s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.748842481 +0000 UTC m=+108.057682282" watchObservedRunningTime="2026-01-26 09:09:34.754354222 +0000 UTC m=+108.063194023" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.775868 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" event={"ID":"835cb68e-1243-4522-bfd0-181378353e51","Type":"ContainerStarted","Data":"a996edbff1468312220e969ebc2527200675e422ccb26ae47fa546443e039d7d"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.778301 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" event={"ID":"07b7e8f0-a421-4fbc-9c0d-aa773210edcd","Type":"ContainerStarted","Data":"0f074f626e397d4d0e1b5b6e4c1aecb86c1a34078c0ce317ae30c6ebcfba7e34"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.782052 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.794003 4872 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bsfdw container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.794074 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" podUID="07b7e8f0-a421-4fbc-9c0d-aa773210edcd" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.795301 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-hvnl5" podStartSLOduration=88.795278617 podStartE2EDuration="1m28.795278617s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.790234418 +0000 UTC m=+108.099074239" watchObservedRunningTime="2026-01-26 09:09:34.795278617 +0000 UTC m=+108.104118418" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.804099 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:34 crc kubenswrapper[4872]: E0126 09:09:34.805514 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:35.305491608 +0000 UTC m=+108.614331409 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.821449 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" event={"ID":"ebeadae0-aa3e-47d1-8068-1c339da55d91","Type":"ContainerStarted","Data":"52c0ef54de1643adfecc5a175def7f25682232038be65e5aaa91842ab5ab8a6b"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.846602 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gfcvb" podStartSLOduration=88.846578756 podStartE2EDuration="1m28.846578756s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.845072234 +0000 UTC m=+108.153912035" watchObservedRunningTime="2026-01-26 09:09:34.846578756 +0000 UTC m=+108.155418557" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.887180 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" event={"ID":"06e81723-7f8b-4dd6-99c6-d1ce1a966b7d","Type":"ContainerStarted","Data":"daa0172472a90211b29b4b3ba4170e5a8e34faf1d99cf29c2f7af4c1db98c8e3"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.897790 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-vvzw4" podStartSLOduration=88.897737811 podStartE2EDuration="1m28.897737811s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.896639631 +0000 UTC m=+108.205479432" watchObservedRunningTime="2026-01-26 09:09:34.897737811 +0000 UTC m=+108.206577622" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.908385 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:34 crc kubenswrapper[4872]: E0126 09:09:34.908777 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:35.408762394 +0000 UTC m=+108.717602195 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.935740 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nwkxd" podStartSLOduration=88.935717864 podStartE2EDuration="1m28.935717864s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.935287123 +0000 UTC m=+108.244126944" watchObservedRunningTime="2026-01-26 09:09:34.935717864 +0000 UTC m=+108.244557665" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.994270 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" event={"ID":"55d2e904-b5e7-4bcf-8c39-1e84e51a2581","Type":"ContainerStarted","Data":"4f342eba80c9256b6adb6f83b61e891400097f0836f4814e6411eea8a47fad38"} Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.994829 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" Jan 26 09:09:34 crc kubenswrapper[4872]: I0126 09:09:34.997733 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-zbn9p" podStartSLOduration=88.997720567 podStartE2EDuration="1m28.997720567s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:34.969087741 +0000 UTC m=+108.277927542" watchObservedRunningTime="2026-01-26 09:09:34.997720567 +0000 UTC m=+108.306560368" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.002370 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5xspp"] Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.014264 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:35 crc kubenswrapper[4872]: E0126 09:09:35.015743 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:35.515704271 +0000 UTC m=+108.824544072 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.016909 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" podStartSLOduration=89.016869783 podStartE2EDuration="1m29.016869783s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:35.013247653 +0000 UTC m=+108.322087454" watchObservedRunningTime="2026-01-26 09:09:35.016869783 +0000 UTC m=+108.325709584" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.057040 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-9tdst" event={"ID":"326f7b91-1e87-4fc0-9913-40f5f93d2cf2","Type":"ContainerStarted","Data":"b83c83874f502c7a7bf95ddc7357668e3b5d7d4d6c63e31b135c5e725d857cd6"} Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.085971 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-zr6z9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.086072 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zr6z9" podUID="c85d43e4-21f0-47fc-ae62-057de42e1935" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.115453 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:35 crc kubenswrapper[4872]: E0126 09:09:35.115911 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:35.615900824 +0000 UTC m=+108.924740625 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.118534 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6gqpg" podStartSLOduration=89.118496594 podStartE2EDuration="1m29.118496594s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:35.056724187 +0000 UTC m=+108.365563988" watchObservedRunningTime="2026-01-26 09:09:35.118496594 +0000 UTC m=+108.427336395" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.147908 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" podStartSLOduration=89.147880351 podStartE2EDuration="1m29.147880351s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:35.113455416 +0000 UTC m=+108.422295207" watchObservedRunningTime="2026-01-26 09:09:35.147880351 +0000 UTC m=+108.456720152" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.148510 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-9tdst" podStartSLOduration=89.148505229 podStartE2EDuration="1m29.148505229s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:35.141461895 +0000 UTC m=+108.450301696" watchObservedRunningTime="2026-01-26 09:09:35.148505229 +0000 UTC m=+108.457345030" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.154083 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-mznfc" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.218047 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:35 crc kubenswrapper[4872]: E0126 09:09:35.219956 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:35.71993119 +0000 UTC m=+109.028770991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.279310 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5z7tb" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.321100 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:35 crc kubenswrapper[4872]: E0126 09:09:35.321593 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:35.821577252 +0000 UTC m=+109.130417053 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.353947 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-01-26 09:04:34 +0000 UTC, rotation deadline is 2026-10-13 06:11:19.713544866 +0000 UTC Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.354015 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6237h1m44.359533523s for next certificate rotation Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.422583 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:35 crc kubenswrapper[4872]: E0126 09:09:35.422921 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:35.922866834 +0000 UTC m=+109.231706635 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.423768 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:35 crc kubenswrapper[4872]: E0126 09:09:35.426753 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:35.92672343 +0000 UTC m=+109.235563221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.531057 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:35 crc kubenswrapper[4872]: E0126 09:09:35.532068 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:36.032044393 +0000 UTC m=+109.340884194 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.550684 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-84qrq"] Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.551884 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.568003 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.572084 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-84qrq"] Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.632389 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.633454 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3132ab11-2629-4089-b4e5-da75a4d83fd9-utilities\") pod \"redhat-marketplace-84qrq\" (UID: \"3132ab11-2629-4089-b4e5-da75a4d83fd9\") " pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.633507 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.633532 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3132ab11-2629-4089-b4e5-da75a4d83fd9-catalog-content\") pod \"redhat-marketplace-84qrq\" (UID: \"3132ab11-2629-4089-b4e5-da75a4d83fd9\") " pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.633589 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p76kx"] Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.633671 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48sxz\" (UniqueName: \"kubernetes.io/projected/3132ab11-2629-4089-b4e5-da75a4d83fd9-kube-api-access-48sxz\") pod \"redhat-marketplace-84qrq\" (UID: \"3132ab11-2629-4089-b4e5-da75a4d83fd9\") " pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:09:35 crc kubenswrapper[4872]: E0126 09:09:35.634793 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:36.134774004 +0000 UTC m=+109.443614005 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.650664 4872 patch_prober.go:28] interesting pod/router-default-5444994796-9tdst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 09:09:35 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Jan 26 09:09:35 crc kubenswrapper[4872]: [+]process-running ok Jan 26 09:09:35 crc kubenswrapper[4872]: healthz check failed Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.650705 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9tdst" podUID="326f7b91-1e87-4fc0-9913-40f5f93d2cf2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.691031 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p5rmt"] Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.735105 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.735321 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3132ab11-2629-4089-b4e5-da75a4d83fd9-utilities\") pod \"redhat-marketplace-84qrq\" (UID: \"3132ab11-2629-4089-b4e5-da75a4d83fd9\") " pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.735382 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3132ab11-2629-4089-b4e5-da75a4d83fd9-catalog-content\") pod \"redhat-marketplace-84qrq\" (UID: \"3132ab11-2629-4089-b4e5-da75a4d83fd9\") " pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.735408 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48sxz\" (UniqueName: \"kubernetes.io/projected/3132ab11-2629-4089-b4e5-da75a4d83fd9-kube-api-access-48sxz\") pod \"redhat-marketplace-84qrq\" (UID: \"3132ab11-2629-4089-b4e5-da75a4d83fd9\") " pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:09:35 crc kubenswrapper[4872]: E0126 09:09:35.735945 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:36.235921942 +0000 UTC m=+109.544761743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.736313 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3132ab11-2629-4089-b4e5-da75a4d83fd9-utilities\") pod \"redhat-marketplace-84qrq\" (UID: \"3132ab11-2629-4089-b4e5-da75a4d83fd9\") " pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.736534 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3132ab11-2629-4089-b4e5-da75a4d83fd9-catalog-content\") pod \"redhat-marketplace-84qrq\" (UID: \"3132ab11-2629-4089-b4e5-da75a4d83fd9\") " pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.777618 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rng9p"] Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.780965 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48sxz\" (UniqueName: \"kubernetes.io/projected/3132ab11-2629-4089-b4e5-da75a4d83fd9-kube-api-access-48sxz\") pod \"redhat-marketplace-84qrq\" (UID: \"3132ab11-2629-4089-b4e5-da75a4d83fd9\") " pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.845935 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:35 crc kubenswrapper[4872]: E0126 09:09:35.846402 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:36.346381136 +0000 UTC m=+109.655220937 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.947067 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:35 crc kubenswrapper[4872]: E0126 09:09:35.947631 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:36.447602056 +0000 UTC m=+109.756441857 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.952096 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.979878 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5qfx7"] Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.990541 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:09:35 crc kubenswrapper[4872]: I0126 09:09:35.993565 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qfx7"] Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.052151 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sm4rl\" (UniqueName: \"kubernetes.io/projected/2d39d8ba-e193-4c19-a535-589ffd091413-kube-api-access-sm4rl\") pod \"redhat-marketplace-5qfx7\" (UID: \"2d39d8ba-e193-4c19-a535-589ffd091413\") " pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.052217 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.052313 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d39d8ba-e193-4c19-a535-589ffd091413-catalog-content\") pod \"redhat-marketplace-5qfx7\" (UID: \"2d39d8ba-e193-4c19-a535-589ffd091413\") " pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.052406 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d39d8ba-e193-4c19-a535-589ffd091413-utilities\") pod \"redhat-marketplace-5qfx7\" (UID: \"2d39d8ba-e193-4c19-a535-589ffd091413\") " pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:09:36 crc kubenswrapper[4872]: E0126 09:09:36.052645 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:36.552624281 +0000 UTC m=+109.861464082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.099640 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lw526" event={"ID":"ea814b88-1c4c-407b-8754-f57e19f8153f","Type":"ContainerStarted","Data":"381970fdac3bb5eef68c238b06c2687eb8872b180aa22a1a7a88146ec24d9fd5"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.112961 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rng9p" event={"ID":"bca70073-3d97-452c-818c-70c8021fe37d","Type":"ContainerStarted","Data":"7381fa6fe6e23850bb1f8bce9429b5f6588df3f83479b198b1f849752a859a7a"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.113020 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rng9p" event={"ID":"bca70073-3d97-452c-818c-70c8021fe37d","Type":"ContainerStarted","Data":"3888844e8ed08ca807359e7b74d94ba3d4424b7f9b29d1e659f162e20c744cd5"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.123988 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.132572 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg" event={"ID":"49b97b45-be77-455a-9a69-f20629af5940","Type":"ContainerStarted","Data":"8fbcd0acd59377de1dc0314fde7e6815fec60fe641ff161d62faaaacb19dfbb7"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.135765 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5rmt" event={"ID":"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e","Type":"ContainerStarted","Data":"63de97ba6862e1a08f8604dd1fbf0c30372ae1b3141813698bbb87ac658c7630"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.135854 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5rmt" event={"ID":"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e","Type":"ContainerStarted","Data":"dbc5d37843ef6b28528a3abc21ad4cf93d711a95a81240dd10563bee59d67ba0"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.140693 4872 generic.go:334] "Generic (PLEG): container finished" podID="7b6eb987-454b-4c50-bb39-ede2e1006808" containerID="433c56cbf0956956461b5d45447e7d80f6370c91b24a8cc388b45707a36112f3" exitCode=0 Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.140769 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xspp" event={"ID":"7b6eb987-454b-4c50-bb39-ede2e1006808","Type":"ContainerDied","Data":"433c56cbf0956956461b5d45447e7d80f6370c91b24a8cc388b45707a36112f3"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.140824 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xspp" event={"ID":"7b6eb987-454b-4c50-bb39-ede2e1006808","Type":"ContainerStarted","Data":"591d31cd59bc4cad588c4b8fa92c06297adbfb9f7f2af9e1689bd75165378942"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.154507 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.154837 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d39d8ba-e193-4c19-a535-589ffd091413-utilities\") pod \"redhat-marketplace-5qfx7\" (UID: \"2d39d8ba-e193-4c19-a535-589ffd091413\") " pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.154925 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sm4rl\" (UniqueName: \"kubernetes.io/projected/2d39d8ba-e193-4c19-a535-589ffd091413-kube-api-access-sm4rl\") pod \"redhat-marketplace-5qfx7\" (UID: \"2d39d8ba-e193-4c19-a535-589ffd091413\") " pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.154981 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d39d8ba-e193-4c19-a535-589ffd091413-catalog-content\") pod \"redhat-marketplace-5qfx7\" (UID: \"2d39d8ba-e193-4c19-a535-589ffd091413\") " pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.155552 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d39d8ba-e193-4c19-a535-589ffd091413-catalog-content\") pod \"redhat-marketplace-5qfx7\" (UID: \"2d39d8ba-e193-4c19-a535-589ffd091413\") " pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:09:36 crc kubenswrapper[4872]: E0126 09:09:36.155639 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:36.655617589 +0000 UTC m=+109.964457390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.155931 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d39d8ba-e193-4c19-a535-589ffd091413-utilities\") pod \"redhat-marketplace-5qfx7\" (UID: \"2d39d8ba-e193-4c19-a535-589ffd091413\") " pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.160054 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8k2xh" event={"ID":"437f0c85-bf20-4e24-9138-cce77c62fd2c","Type":"ContainerStarted","Data":"4dcd2e52c71c3ce8ab33cb52514c53248262f25ef57393aee4423279327ad525"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.160119 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-8k2xh" event={"ID":"437f0c85-bf20-4e24-9138-cce77c62fd2c","Type":"ContainerStarted","Data":"5b6a43315e57ee41da1e8ccfaa4deb910a6ea4eedda9ae82bd9d4a2c71b29975"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.170520 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cmb5q" event={"ID":"7e3a84de-b4dd-4fc0-a279-9e2a9a616216","Type":"ContainerStarted","Data":"7bbae1daf30afeb8885db2ac2ab1edc7a98bee39e13dba4bafd4852963f8fead"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.170934 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.179591 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" event={"ID":"7975e724-e9f6-4a68-8779-f894ac687f74","Type":"ContainerStarted","Data":"9b78fe2f0337080f2173940306e62f31f4fc8d5d6f60c99c5d6963f2768d58f9"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.183547 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p76kx" event={"ID":"c963ddba-b620-46b4-83af-dfbaa96bbd03","Type":"ContainerStarted","Data":"1f9ac8ebee536943cb2adc8ff1454826058ea508b69dd0568adcd740c1a43cf6"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.183605 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p76kx" event={"ID":"c963ddba-b620-46b4-83af-dfbaa96bbd03","Type":"ContainerStarted","Data":"b2a83f96f6f7d8a3a38b1aac0e94d08e82a8e2bc7bec8afaba3890b50f369993"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.188269 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sm4rl\" (UniqueName: \"kubernetes.io/projected/2d39d8ba-e193-4c19-a535-589ffd091413-kube-api-access-sm4rl\") pod \"redhat-marketplace-5qfx7\" (UID: \"2d39d8ba-e193-4c19-a535-589ffd091413\") " pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.195931 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.197737 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" event={"ID":"55d2e904-b5e7-4bcf-8c39-1e84e51a2581","Type":"ContainerStarted","Data":"84fe9a65827814e63733ef9460b3bd147a4a243f85bd5be27f7c4eb07929a5a3"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.219185 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wfcsg" podStartSLOduration=90.219159425 podStartE2EDuration="1m30.219159425s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:36.218611999 +0000 UTC m=+109.527451800" watchObservedRunningTime="2026-01-26 09:09:36.219159425 +0000 UTC m=+109.527999226" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.223478 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pplqc" event={"ID":"1bd8496f-6499-4cf1-894a-37f68df536f4","Type":"ContainerStarted","Data":"91ff718c6f65d8748db574f56dd448578d47b98357163f6f89e848046adfca5c"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.241359 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" event={"ID":"ef55e8dc-7dfd-47e4-aaf2-7399eb9603d6","Type":"ContainerStarted","Data":"f8874626b495ec3f50ab341debe3c0911ed52ecf39dc804993ffcf8248bb5af0"} Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.242479 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-zr6z9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.242540 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zr6z9" podUID="c85d43e4-21f0-47fc-ae62-057de42e1935" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.262079 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.272129 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:36 crc kubenswrapper[4872]: E0126 09:09:36.274421 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:36.774392581 +0000 UTC m=+110.083232382 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.329668 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-4hhdc" podStartSLOduration=90.329649809 podStartE2EDuration="1m30.329649809s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:36.329350721 +0000 UTC m=+109.638190522" watchObservedRunningTime="2026-01-26 09:09:36.329649809 +0000 UTC m=+109.638489610" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.331360 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-ntnp2" podStartSLOduration=90.331351826 podStartE2EDuration="1m30.331351826s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:36.303362947 +0000 UTC m=+109.612202748" watchObservedRunningTime="2026-01-26 09:09:36.331351826 +0000 UTC m=+109.640191627" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.373643 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:36 crc kubenswrapper[4872]: E0126 09:09:36.376515 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:36.876492456 +0000 UTC m=+110.185332267 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.418956 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-cmb5q" podStartSLOduration=10.418930271 podStartE2EDuration="10.418930271s" podCreationTimestamp="2026-01-26 09:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:36.386533021 +0000 UTC m=+109.695372822" watchObservedRunningTime="2026-01-26 09:09:36.418930271 +0000 UTC m=+109.727770072" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.465930 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-8k2xh" podStartSLOduration=90.465904412 podStartE2EDuration="1m30.465904412s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:36.419477257 +0000 UTC m=+109.728317058" watchObservedRunningTime="2026-01-26 09:09:36.465904412 +0000 UTC m=+109.774744213" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.478104 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:36 crc kubenswrapper[4872]: E0126 09:09:36.478656 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:36.978635791 +0000 UTC m=+110.287475602 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.581363 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:36 crc kubenswrapper[4872]: E0126 09:09:36.581770 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:37.081751234 +0000 UTC m=+110.390591035 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.637017 4872 patch_prober.go:28] interesting pod/router-default-5444994796-9tdst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 09:09:36 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Jan 26 09:09:36 crc kubenswrapper[4872]: [+]process-running ok Jan 26 09:09:36 crc kubenswrapper[4872]: healthz check failed Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.637070 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9tdst" podUID="326f7b91-1e87-4fc0-9913-40f5f93d2cf2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.684035 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:36 crc kubenswrapper[4872]: E0126 09:09:36.684502 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:37.184484345 +0000 UTC m=+110.493324146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.745631 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2cwfr"] Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.748621 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.752612 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.775916 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-84qrq"] Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.785409 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.785858 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eedb2ebd-5684-464d-bc70-fba04fb6edbd-utilities\") pod \"redhat-operators-2cwfr\" (UID: \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\") " pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.785939 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdf9w\" (UniqueName: \"kubernetes.io/projected/eedb2ebd-5684-464d-bc70-fba04fb6edbd-kube-api-access-qdf9w\") pod \"redhat-operators-2cwfr\" (UID: \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\") " pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.786009 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eedb2ebd-5684-464d-bc70-fba04fb6edbd-catalog-content\") pod \"redhat-operators-2cwfr\" (UID: \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\") " pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:09:36 crc kubenswrapper[4872]: E0126 09:09:36.786141 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:37.286119456 +0000 UTC m=+110.594959257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.824970 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2cwfr"] Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.887437 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.887494 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdf9w\" (UniqueName: \"kubernetes.io/projected/eedb2ebd-5684-464d-bc70-fba04fb6edbd-kube-api-access-qdf9w\") pod \"redhat-operators-2cwfr\" (UID: \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\") " pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.887588 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eedb2ebd-5684-464d-bc70-fba04fb6edbd-catalog-content\") pod \"redhat-operators-2cwfr\" (UID: \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\") " pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.887637 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eedb2ebd-5684-464d-bc70-fba04fb6edbd-utilities\") pod \"redhat-operators-2cwfr\" (UID: \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\") " pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:09:36 crc kubenswrapper[4872]: E0126 09:09:36.888027 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:37.388001714 +0000 UTC m=+110.696841515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.888216 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eedb2ebd-5684-464d-bc70-fba04fb6edbd-utilities\") pod \"redhat-operators-2cwfr\" (UID: \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\") " pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.888513 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eedb2ebd-5684-464d-bc70-fba04fb6edbd-catalog-content\") pod \"redhat-operators-2cwfr\" (UID: \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\") " pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.939032 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdf9w\" (UniqueName: \"kubernetes.io/projected/eedb2ebd-5684-464d-bc70-fba04fb6edbd-kube-api-access-qdf9w\") pod \"redhat-operators-2cwfr\" (UID: \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\") " pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.966504 4872 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.989354 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:36 crc kubenswrapper[4872]: E0126 09:09:36.989617 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:37.489565674 +0000 UTC m=+110.798405475 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:36 crc kubenswrapper[4872]: I0126 09:09:36.989995 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:36 crc kubenswrapper[4872]: E0126 09:09:36.990501 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:37.490484139 +0000 UTC m=+110.799323930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.082343 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qfx7"] Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.083646 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.091543 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:37 crc kubenswrapper[4872]: E0126 09:09:37.092570 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:37.592548902 +0000 UTC m=+110.901388703 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.157640 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-96hdr"] Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.158987 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.199071 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.199128 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcn9t\" (UniqueName: \"kubernetes.io/projected/f8525631-7064-4a4e-93ce-a5a5e0f77274-kube-api-access-xcn9t\") pod \"redhat-operators-96hdr\" (UID: \"f8525631-7064-4a4e-93ce-a5a5e0f77274\") " pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.199161 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8525631-7064-4a4e-93ce-a5a5e0f77274-catalog-content\") pod \"redhat-operators-96hdr\" (UID: \"f8525631-7064-4a4e-93ce-a5a5e0f77274\") " pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.199230 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8525631-7064-4a4e-93ce-a5a5e0f77274-utilities\") pod \"redhat-operators-96hdr\" (UID: \"f8525631-7064-4a4e-93ce-a5a5e0f77274\") " pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:09:37 crc kubenswrapper[4872]: E0126 09:09:37.199937 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:37.699921711 +0000 UTC m=+111.008761512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.235791 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-96hdr"] Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.281611 4872 generic.go:334] "Generic (PLEG): container finished" podID="c963ddba-b620-46b4-83af-dfbaa96bbd03" containerID="1f9ac8ebee536943cb2adc8ff1454826058ea508b69dd0568adcd740c1a43cf6" exitCode=0 Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.281822 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p76kx" event={"ID":"c963ddba-b620-46b4-83af-dfbaa96bbd03","Type":"ContainerDied","Data":"1f9ac8ebee536943cb2adc8ff1454826058ea508b69dd0568adcd740c1a43cf6"} Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.300321 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.300641 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8525631-7064-4a4e-93ce-a5a5e0f77274-utilities\") pod \"redhat-operators-96hdr\" (UID: \"f8525631-7064-4a4e-93ce-a5a5e0f77274\") " pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.300823 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcn9t\" (UniqueName: \"kubernetes.io/projected/f8525631-7064-4a4e-93ce-a5a5e0f77274-kube-api-access-xcn9t\") pod \"redhat-operators-96hdr\" (UID: \"f8525631-7064-4a4e-93ce-a5a5e0f77274\") " pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.300847 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8525631-7064-4a4e-93ce-a5a5e0f77274-catalog-content\") pod \"redhat-operators-96hdr\" (UID: \"f8525631-7064-4a4e-93ce-a5a5e0f77274\") " pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:09:37 crc kubenswrapper[4872]: E0126 09:09:37.300932 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-01-26 09:09:37.800893225 +0000 UTC m=+111.109733226 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.301765 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8525631-7064-4a4e-93ce-a5a5e0f77274-catalog-content\") pod \"redhat-operators-96hdr\" (UID: \"f8525631-7064-4a4e-93ce-a5a5e0f77274\") " pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.302190 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8525631-7064-4a4e-93ce-a5a5e0f77274-utilities\") pod \"redhat-operators-96hdr\" (UID: \"f8525631-7064-4a4e-93ce-a5a5e0f77274\") " pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.349581 4872 generic.go:334] "Generic (PLEG): container finished" podID="90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" containerID="63de97ba6862e1a08f8604dd1fbf0c30372ae1b3141813698bbb87ac658c7630" exitCode=0 Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.349684 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5rmt" event={"ID":"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e","Type":"ContainerDied","Data":"63de97ba6862e1a08f8604dd1fbf0c30372ae1b3141813698bbb87ac658c7630"} Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.386937 4872 generic.go:334] "Generic (PLEG): container finished" podID="3132ab11-2629-4089-b4e5-da75a4d83fd9" containerID="19e7a143bdb0f3e88e3be99e0b7f23e6960a8df3b681cd2e00b741f197ef2695" exitCode=0 Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.387035 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84qrq" event={"ID":"3132ab11-2629-4089-b4e5-da75a4d83fd9","Type":"ContainerDied","Data":"19e7a143bdb0f3e88e3be99e0b7f23e6960a8df3b681cd2e00b741f197ef2695"} Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.387065 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84qrq" event={"ID":"3132ab11-2629-4089-b4e5-da75a4d83fd9","Type":"ContainerStarted","Data":"3159a5365b4a224d049d3d2c32bc0ffaaa1fb28b6f22562b6d76a5b9610b1be5"} Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.387410 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcn9t\" (UniqueName: \"kubernetes.io/projected/f8525631-7064-4a4e-93ce-a5a5e0f77274-kube-api-access-xcn9t\") pod \"redhat-operators-96hdr\" (UID: \"f8525631-7064-4a4e-93ce-a5a5e0f77274\") " pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.393132 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qfx7" event={"ID":"2d39d8ba-e193-4c19-a535-589ffd091413","Type":"ContainerStarted","Data":"43b03334a8ba72b9d7d3e396fc432fb4d859a01f1c617018396074aa085148d7"} Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.403858 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:37 crc kubenswrapper[4872]: E0126 09:09:37.404648 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-01-26 09:09:37.904625293 +0000 UTC m=+111.213465094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-p975q" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.408525 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lw526" event={"ID":"ea814b88-1c4c-407b-8754-f57e19f8153f","Type":"ContainerStarted","Data":"07520a2a6534ab157737505e8a870fd4128e89ddfc756dfafb9b77b008bfeadd"} Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.415792 4872 generic.go:334] "Generic (PLEG): container finished" podID="bca70073-3d97-452c-818c-70c8021fe37d" containerID="7381fa6fe6e23850bb1f8bce9429b5f6588df3f83479b198b1f849752a859a7a" exitCode=0 Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.416792 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rng9p" event={"ID":"bca70073-3d97-452c-818c-70c8021fe37d","Type":"ContainerDied","Data":"7381fa6fe6e23850bb1f8bce9429b5f6588df3f83479b198b1f849752a859a7a"} Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.450401 4872 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-01-26T09:09:36.966551742Z","Handler":null,"Name":""} Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.460065 4872 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.460118 4872 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.504461 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.505939 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.593773 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.622127 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.627213 4872 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.627308 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.636310 4872 patch_prober.go:28] interesting pod/router-default-5444994796-9tdst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 09:09:37 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Jan 26 09:09:37 crc kubenswrapper[4872]: [+]process-running ok Jan 26 09:09:37 crc kubenswrapper[4872]: healthz check failed Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.636416 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9tdst" podUID="326f7b91-1e87-4fc0-9913-40f5f93d2cf2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.637412 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2cwfr"] Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.683447 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-p975q\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.826819 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.842536 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-82cml" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.894028 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.895144 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.922498 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.922788 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.923443 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.925696 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.936920 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8bce1b95-50f4-4d87-b7d4-6cae890bfef2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8bce1b95-50f4-4d87-b7d4-6cae890bfef2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.937203 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8bce1b95-50f4-4d87-b7d4-6cae890bfef2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8bce1b95-50f4-4d87-b7d4-6cae890bfef2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 09:09:37 crc kubenswrapper[4872]: I0126 09:09:37.977603 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-96hdr"] Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.040146 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8bce1b95-50f4-4d87-b7d4-6cae890bfef2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8bce1b95-50f4-4d87-b7d4-6cae890bfef2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.040245 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8bce1b95-50f4-4d87-b7d4-6cae890bfef2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8bce1b95-50f4-4d87-b7d4-6cae890bfef2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.040358 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8bce1b95-50f4-4d87-b7d4-6cae890bfef2-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8bce1b95-50f4-4d87-b7d4-6cae890bfef2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.120880 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8bce1b95-50f4-4d87-b7d4-6cae890bfef2-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8bce1b95-50f4-4d87-b7d4-6cae890bfef2\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.374869 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.470070 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p975q"] Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.511828 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lw526" event={"ID":"ea814b88-1c4c-407b-8754-f57e19f8153f","Type":"ContainerStarted","Data":"0b6365346aa48d8bbd40e1f1364e81d4d37c5fe1cf56ebbb6c104a68a2f701fc"} Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.521455 4872 generic.go:334] "Generic (PLEG): container finished" podID="f8525631-7064-4a4e-93ce-a5a5e0f77274" containerID="713be9d4ef286298e87284c52944fc06c70c5bf43723456d532461662d378ea6" exitCode=0 Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.522315 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96hdr" event={"ID":"f8525631-7064-4a4e-93ce-a5a5e0f77274","Type":"ContainerDied","Data":"713be9d4ef286298e87284c52944fc06c70c5bf43723456d532461662d378ea6"} Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.522339 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96hdr" event={"ID":"f8525631-7064-4a4e-93ce-a5a5e0f77274","Type":"ContainerStarted","Data":"befe57ef566e35a3767e5693dddbe843d8b0ea10ea0e41ba7a2178633e7f3e40"} Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.583030 4872 generic.go:334] "Generic (PLEG): container finished" podID="f2c67438-c0c7-48b8-a3df-21669ccb1602" containerID="76664482753c3c2630369c1de9f8fe161bea6d4557a168d0d81b98f6c2e95d4c" exitCode=0 Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.583130 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" event={"ID":"f2c67438-c0c7-48b8-a3df-21669ccb1602","Type":"ContainerDied","Data":"76664482753c3c2630369c1de9f8fe161bea6d4557a168d0d81b98f6c2e95d4c"} Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.604769 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qfx7" event={"ID":"2d39d8ba-e193-4c19-a535-589ffd091413","Type":"ContainerDied","Data":"0c19873dc62882d3c604e85ef7e900b7d9add7cf21d701109258e9bd690f7ea9"} Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.604576 4872 generic.go:334] "Generic (PLEG): container finished" podID="2d39d8ba-e193-4c19-a535-589ffd091413" containerID="0c19873dc62882d3c604e85ef7e900b7d9add7cf21d701109258e9bd690f7ea9" exitCode=0 Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.608029 4872 generic.go:334] "Generic (PLEG): container finished" podID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" containerID="f10dba94e9487ea0c17a2c4068795ef39906b3cbbea831beb19e875132123eae" exitCode=0 Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.608827 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cwfr" event={"ID":"eedb2ebd-5684-464d-bc70-fba04fb6edbd","Type":"ContainerDied","Data":"f10dba94e9487ea0c17a2c4068795ef39906b3cbbea831beb19e875132123eae"} Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.608916 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cwfr" event={"ID":"eedb2ebd-5684-464d-bc70-fba04fb6edbd","Type":"ContainerStarted","Data":"4a0543754d2361565b10a92631d8b96118106db0bea4268ab7601df3d8990d4e"} Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.643847 4872 patch_prober.go:28] interesting pod/router-default-5444994796-9tdst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 09:09:38 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Jan 26 09:09:38 crc kubenswrapper[4872]: [+]process-running ok Jan 26 09:09:38 crc kubenswrapper[4872]: healthz check failed Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.643898 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9tdst" podUID="326f7b91-1e87-4fc0-9913-40f5f93d2cf2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.677367 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.682348 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.691001 4872 patch_prober.go:28] interesting pod/console-f9d7485db-ztd4t container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.691052 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ztd4t" podUID="9d625f91-d4de-4687-bdbe-8ff985d7e406" containerName="console" probeResult="failure" output="Get \"https://10.217.0.24:8443/health\": dial tcp 10.217.0.24:8443: connect: connection refused" Jan 26 09:09:38 crc kubenswrapper[4872]: I0126 09:09:38.980653 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Jan 26 09:09:39 crc kubenswrapper[4872]: W0126 09:09:39.043010 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod8bce1b95_50f4_4d87_b7d4_6cae890bfef2.slice/crio-9cdc1f11c348d41e03c65a298078cf3635b7b28fc20c76755a0372a81741975a WatchSource:0}: Error finding container 9cdc1f11c348d41e03c65a298078cf3635b7b28fc20c76755a0372a81741975a: Status 404 returned error can't find the container with id 9cdc1f11c348d41e03c65a298078cf3635b7b28fc20c76755a0372a81741975a Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.200446 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.573009 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-zr6z9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.573103 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zr6z9" podUID="c85d43e4-21f0-47fc-ae62-057de42e1935" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.573231 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-zr6z9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.573361 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-zr6z9" podUID="c85d43e4-21f0-47fc-ae62-057de42e1935" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.631823 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.639110 4872 patch_prober.go:28] interesting pod/router-default-5444994796-9tdst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 09:09:39 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Jan 26 09:09:39 crc kubenswrapper[4872]: [+]process-running ok Jan 26 09:09:39 crc kubenswrapper[4872]: healthz check failed Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.639195 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9tdst" podUID="326f7b91-1e87-4fc0-9913-40f5f93d2cf2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.644399 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" event={"ID":"9aa4a2c3-14ef-42ee-a4cb-261166d7350d","Type":"ContainerStarted","Data":"f82d544f596d31854f94c8c1eaffeeffc99cb646823caa0f58a123c61e06ed93"} Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.644450 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" event={"ID":"9aa4a2c3-14ef-42ee-a4cb-261166d7350d","Type":"ContainerStarted","Data":"8bcb06d1756f46daf488c8f81628f94a3382dd5a2a51572dce0912d2e353b845"} Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.644957 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.700831 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lw526" event={"ID":"ea814b88-1c4c-407b-8754-f57e19f8153f","Type":"ContainerStarted","Data":"fc06d5b5611779c117940e5da7ddd0e048320dd7655fbc4ff2e609ff8ad9c2af"} Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.708055 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8bce1b95-50f4-4d87-b7d4-6cae890bfef2","Type":"ContainerStarted","Data":"9cdc1f11c348d41e03c65a298078cf3635b7b28fc20c76755a0372a81741975a"} Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.731267 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" podStartSLOduration=93.731241545 podStartE2EDuration="1m33.731241545s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:39.67282128 +0000 UTC m=+112.981661091" watchObservedRunningTime="2026-01-26 09:09:39.731241545 +0000 UTC m=+113.040081346" Jan 26 09:09:39 crc kubenswrapper[4872]: I0126 09:09:39.739890 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-lw526" podStartSLOduration=13.739861801 podStartE2EDuration="13.739861801s" podCreationTimestamp="2026-01-26 09:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:09:39.739206424 +0000 UTC m=+113.048046225" watchObservedRunningTime="2026-01-26 09:09:39.739861801 +0000 UTC m=+113.048701652" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.205890 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.229294 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2c67438-c0c7-48b8-a3df-21669ccb1602-config-volume\") pod \"f2c67438-c0c7-48b8-a3df-21669ccb1602\" (UID: \"f2c67438-c0c7-48b8-a3df-21669ccb1602\") " Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.229381 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2c67438-c0c7-48b8-a3df-21669ccb1602-secret-volume\") pod \"f2c67438-c0c7-48b8-a3df-21669ccb1602\" (UID: \"f2c67438-c0c7-48b8-a3df-21669ccb1602\") " Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.229413 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vpt9\" (UniqueName: \"kubernetes.io/projected/f2c67438-c0c7-48b8-a3df-21669ccb1602-kube-api-access-8vpt9\") pod \"f2c67438-c0c7-48b8-a3df-21669ccb1602\" (UID: \"f2c67438-c0c7-48b8-a3df-21669ccb1602\") " Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.231460 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2c67438-c0c7-48b8-a3df-21669ccb1602-config-volume" (OuterVolumeSpecName: "config-volume") pod "f2c67438-c0c7-48b8-a3df-21669ccb1602" (UID: "f2c67438-c0c7-48b8-a3df-21669ccb1602"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.255147 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2c67438-c0c7-48b8-a3df-21669ccb1602-kube-api-access-8vpt9" (OuterVolumeSpecName: "kube-api-access-8vpt9") pod "f2c67438-c0c7-48b8-a3df-21669ccb1602" (UID: "f2c67438-c0c7-48b8-a3df-21669ccb1602"). InnerVolumeSpecName "kube-api-access-8vpt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.255332 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2c67438-c0c7-48b8-a3df-21669ccb1602-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f2c67438-c0c7-48b8-a3df-21669ccb1602" (UID: "f2c67438-c0c7-48b8-a3df-21669ccb1602"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.332912 4872 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f2c67438-c0c7-48b8-a3df-21669ccb1602-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.332958 4872 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f2c67438-c0c7-48b8-a3df-21669ccb1602-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.332969 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vpt9\" (UniqueName: \"kubernetes.io/projected/f2c67438-c0c7-48b8-a3df-21669ccb1602-kube-api-access-8vpt9\") on node \"crc\" DevicePath \"\"" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.526566 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 26 09:09:40 crc kubenswrapper[4872]: E0126 09:09:40.526910 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2c67438-c0c7-48b8-a3df-21669ccb1602" containerName="collect-profiles" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.526933 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2c67438-c0c7-48b8-a3df-21669ccb1602" containerName="collect-profiles" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.527095 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2c67438-c0c7-48b8-a3df-21669ccb1602" containerName="collect-profiles" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.527608 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.533360 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.533549 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.540674 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.637863 4872 patch_prober.go:28] interesting pod/router-default-5444994796-9tdst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 09:09:40 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Jan 26 09:09:40 crc kubenswrapper[4872]: [+]process-running ok Jan 26 09:09:40 crc kubenswrapper[4872]: healthz check failed Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.637960 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9tdst" podUID="326f7b91-1e87-4fc0-9913-40f5f93d2cf2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.638434 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17f3f22d-7f02-4cec-951a-2b05bd144293-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"17f3f22d-7f02-4cec-951a-2b05bd144293\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.638491 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/17f3f22d-7f02-4cec-951a-2b05bd144293-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"17f3f22d-7f02-4cec-951a-2b05bd144293\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.743880 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17f3f22d-7f02-4cec-951a-2b05bd144293-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"17f3f22d-7f02-4cec-951a-2b05bd144293\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.743976 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/17f3f22d-7f02-4cec-951a-2b05bd144293-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"17f3f22d-7f02-4cec-951a-2b05bd144293\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.744130 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/17f3f22d-7f02-4cec-951a-2b05bd144293-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"17f3f22d-7f02-4cec-951a-2b05bd144293\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.770454 4872 generic.go:334] "Generic (PLEG): container finished" podID="8bce1b95-50f4-4d87-b7d4-6cae890bfef2" containerID="98501fdb527bda8a413d2c600edb92da3ca5925bc4f43ddf9a84d18bc3dd3f2f" exitCode=0 Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.770634 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8bce1b95-50f4-4d87-b7d4-6cae890bfef2","Type":"ContainerDied","Data":"98501fdb527bda8a413d2c600edb92da3ca5925bc4f43ddf9a84d18bc3dd3f2f"} Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.770766 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17f3f22d-7f02-4cec-951a-2b05bd144293-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"17f3f22d-7f02-4cec-951a-2b05bd144293\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.776946 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.781574 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs" event={"ID":"f2c67438-c0c7-48b8-a3df-21669ccb1602","Type":"ContainerDied","Data":"77d3c90487255fd31a97c328df74ad31f693023f2b3e0048e4861275ef2b0313"} Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.781676 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77d3c90487255fd31a97c328df74ad31f693023f2b3e0048e4861275ef2b0313" Jan 26 09:09:40 crc kubenswrapper[4872]: I0126 09:09:40.869572 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 09:09:41 crc kubenswrapper[4872]: I0126 09:09:41.252990 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Jan 26 09:09:41 crc kubenswrapper[4872]: W0126 09:09:41.284777 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod17f3f22d_7f02_4cec_951a_2b05bd144293.slice/crio-a2b2b3b1623c81a9a4800085e6a9150fa70474c3afe0515c787ecf2e545db92e WatchSource:0}: Error finding container a2b2b3b1623c81a9a4800085e6a9150fa70474c3afe0515c787ecf2e545db92e: Status 404 returned error can't find the container with id a2b2b3b1623c81a9a4800085e6a9150fa70474c3afe0515c787ecf2e545db92e Jan 26 09:09:41 crc kubenswrapper[4872]: I0126 09:09:41.635631 4872 patch_prober.go:28] interesting pod/router-default-5444994796-9tdst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 09:09:41 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Jan 26 09:09:41 crc kubenswrapper[4872]: [+]process-running ok Jan 26 09:09:41 crc kubenswrapper[4872]: healthz check failed Jan 26 09:09:41 crc kubenswrapper[4872]: I0126 09:09:41.635705 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9tdst" podUID="326f7b91-1e87-4fc0-9913-40f5f93d2cf2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 09:09:41 crc kubenswrapper[4872]: I0126 09:09:41.801120 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"17f3f22d-7f02-4cec-951a-2b05bd144293","Type":"ContainerStarted","Data":"a2b2b3b1623c81a9a4800085e6a9150fa70474c3afe0515c787ecf2e545db92e"} Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.180264 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.281029 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8bce1b95-50f4-4d87-b7d4-6cae890bfef2-kube-api-access\") pod \"8bce1b95-50f4-4d87-b7d4-6cae890bfef2\" (UID: \"8bce1b95-50f4-4d87-b7d4-6cae890bfef2\") " Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.281222 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8bce1b95-50f4-4d87-b7d4-6cae890bfef2-kubelet-dir\") pod \"8bce1b95-50f4-4d87-b7d4-6cae890bfef2\" (UID: \"8bce1b95-50f4-4d87-b7d4-6cae890bfef2\") " Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.281588 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8bce1b95-50f4-4d87-b7d4-6cae890bfef2-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "8bce1b95-50f4-4d87-b7d4-6cae890bfef2" (UID: "8bce1b95-50f4-4d87-b7d4-6cae890bfef2"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.281781 4872 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8bce1b95-50f4-4d87-b7d4-6cae890bfef2-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.312499 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bce1b95-50f4-4d87-b7d4-6cae890bfef2-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "8bce1b95-50f4-4d87-b7d4-6cae890bfef2" (UID: "8bce1b95-50f4-4d87-b7d4-6cae890bfef2"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.384454 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8bce1b95-50f4-4d87-b7d4-6cae890bfef2-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.642089 4872 patch_prober.go:28] interesting pod/router-default-5444994796-9tdst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 09:09:42 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Jan 26 09:09:42 crc kubenswrapper[4872]: [+]process-running ok Jan 26 09:09:42 crc kubenswrapper[4872]: healthz check failed Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.642187 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9tdst" podUID="326f7b91-1e87-4fc0-9913-40f5f93d2cf2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.835436 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.836701 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8bce1b95-50f4-4d87-b7d4-6cae890bfef2","Type":"ContainerDied","Data":"9cdc1f11c348d41e03c65a298078cf3635b7b28fc20c76755a0372a81741975a"} Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.836749 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cdc1f11c348d41e03c65a298078cf3635b7b28fc20c76755a0372a81741975a" Jan 26 09:09:42 crc kubenswrapper[4872]: I0126 09:09:42.846203 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"17f3f22d-7f02-4cec-951a-2b05bd144293","Type":"ContainerStarted","Data":"62521866abc658306a2ba06c5854d2528f12cf7310818170aefcb6d1710574fc"} Jan 26 09:09:43 crc kubenswrapper[4872]: I0126 09:09:43.634235 4872 patch_prober.go:28] interesting pod/router-default-5444994796-9tdst container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Jan 26 09:09:43 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Jan 26 09:09:43 crc kubenswrapper[4872]: [+]process-running ok Jan 26 09:09:43 crc kubenswrapper[4872]: healthz check failed Jan 26 09:09:43 crc kubenswrapper[4872]: I0126 09:09:43.634309 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-9tdst" podUID="326f7b91-1e87-4fc0-9913-40f5f93d2cf2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Jan 26 09:09:43 crc kubenswrapper[4872]: I0126 09:09:43.858846 4872 generic.go:334] "Generic (PLEG): container finished" podID="17f3f22d-7f02-4cec-951a-2b05bd144293" containerID="62521866abc658306a2ba06c5854d2528f12cf7310818170aefcb6d1710574fc" exitCode=0 Jan 26 09:09:43 crc kubenswrapper[4872]: I0126 09:09:43.858933 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"17f3f22d-7f02-4cec-951a-2b05bd144293","Type":"ContainerDied","Data":"62521866abc658306a2ba06c5854d2528f12cf7310818170aefcb6d1710574fc"} Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.221237 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.347269 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17f3f22d-7f02-4cec-951a-2b05bd144293-kube-api-access\") pod \"17f3f22d-7f02-4cec-951a-2b05bd144293\" (UID: \"17f3f22d-7f02-4cec-951a-2b05bd144293\") " Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.347322 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/17f3f22d-7f02-4cec-951a-2b05bd144293-kubelet-dir\") pod \"17f3f22d-7f02-4cec-951a-2b05bd144293\" (UID: \"17f3f22d-7f02-4cec-951a-2b05bd144293\") " Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.347904 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/17f3f22d-7f02-4cec-951a-2b05bd144293-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "17f3f22d-7f02-4cec-951a-2b05bd144293" (UID: "17f3f22d-7f02-4cec-951a-2b05bd144293"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.355189 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17f3f22d-7f02-4cec-951a-2b05bd144293-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "17f3f22d-7f02-4cec-951a-2b05bd144293" (UID: "17f3f22d-7f02-4cec-951a-2b05bd144293"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.449632 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/17f3f22d-7f02-4cec-951a-2b05bd144293-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.449670 4872 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/17f3f22d-7f02-4cec-951a-2b05bd144293-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.469532 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-cmb5q" Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.637440 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.648533 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-9tdst" Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.902555 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.903685 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"17f3f22d-7f02-4cec-951a-2b05bd144293","Type":"ContainerDied","Data":"a2b2b3b1623c81a9a4800085e6a9150fa70474c3afe0515c787ecf2e545db92e"} Jan 26 09:09:44 crc kubenswrapper[4872]: I0126 09:09:44.903714 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2b2b3b1623c81a9a4800085e6a9150fa70474c3afe0515c787ecf2e545db92e" Jan 26 09:09:48 crc kubenswrapper[4872]: I0126 09:09:48.682611 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:48 crc kubenswrapper[4872]: I0126 09:09:48.688822 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:09:49 crc kubenswrapper[4872]: I0126 09:09:49.587934 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-zr6z9" Jan 26 09:09:53 crc kubenswrapper[4872]: I0126 09:09:53.150644 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:09:57 crc kubenswrapper[4872]: I0126 09:09:57.932074 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:10:09 crc kubenswrapper[4872]: I0126 09:10:09.618567 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-dcs6q" Jan 26 09:10:10 crc kubenswrapper[4872]: I0126 09:10:10.227958 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:10:10 crc kubenswrapper[4872]: I0126 09:10:10.228121 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:10:11 crc kubenswrapper[4872]: E0126 09:10:11.498691 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 26 09:10:11 crc kubenswrapper[4872]: E0126 09:10:11.499349 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sm4rl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-5qfx7_openshift-marketplace(2d39d8ba-e193-4c19-a535-589ffd091413): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 09:10:11 crc kubenswrapper[4872]: E0126 09:10:11.500782 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-5qfx7" podUID="2d39d8ba-e193-4c19-a535-589ffd091413" Jan 26 09:10:11 crc kubenswrapper[4872]: E0126 09:10:11.501817 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Jan 26 09:10:11 crc kubenswrapper[4872]: E0126 09:10:11.501901 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-48sxz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-84qrq_openshift-marketplace(3132ab11-2629-4089-b4e5-da75a4d83fd9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 09:10:11 crc kubenswrapper[4872]: E0126 09:10:11.503061 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-84qrq" podUID="3132ab11-2629-4089-b4e5-da75a4d83fd9" Jan 26 09:10:13 crc kubenswrapper[4872]: E0126 09:10:13.008597 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-84qrq" podUID="3132ab11-2629-4089-b4e5-da75a4d83fd9" Jan 26 09:10:13 crc kubenswrapper[4872]: E0126 09:10:13.008901 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-5qfx7" podUID="2d39d8ba-e193-4c19-a535-589ffd091413" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.119689 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.120240 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.120321 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.120360 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.122207 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.122846 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.123491 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.131694 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.132615 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.137498 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.144302 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.145492 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:10:13 crc kubenswrapper[4872]: E0126 09:10:13.240209 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Jan 26 09:10:13 crc kubenswrapper[4872]: E0126 09:10:13.240398 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-br9rq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-5xspp_openshift-marketplace(7b6eb987-454b-4c50-bb39-ede2e1006808): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Jan 26 09:10:13 crc kubenswrapper[4872]: E0126 09:10:13.242702 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-5xspp" podUID="7b6eb987-454b-4c50-bb39-ede2e1006808" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.403502 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.413241 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Jan 26 09:10:13 crc kubenswrapper[4872]: I0126 09:10:13.426485 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Jan 26 09:10:14 crc kubenswrapper[4872]: W0126 09:10:14.007331 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-f257f87d723194980907fa8e5aa1966e5008c0d5b9a8292916a0960c8619bbc2 WatchSource:0}: Error finding container f257f87d723194980907fa8e5aa1966e5008c0d5b9a8292916a0960c8619bbc2: Status 404 returned error can't find the container with id f257f87d723194980907fa8e5aa1966e5008c0d5b9a8292916a0960c8619bbc2 Jan 26 09:10:14 crc kubenswrapper[4872]: W0126 09:10:14.008467 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-226bbdcc2d9ab662e3951eed6cf67e4d3bd07a841fa637f7455f6dded07c7453 WatchSource:0}: Error finding container 226bbdcc2d9ab662e3951eed6cf67e4d3bd07a841fa637f7455f6dded07c7453: Status 404 returned error can't find the container with id 226bbdcc2d9ab662e3951eed6cf67e4d3bd07a841fa637f7455f6dded07c7453 Jan 26 09:10:14 crc kubenswrapper[4872]: I0126 09:10:14.123408 4872 generic.go:334] "Generic (PLEG): container finished" podID="90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" containerID="cde797e3deb4926f06c3da00432d83090b3025711742fca2c560fc092f42b65b" exitCode=0 Jan 26 09:10:14 crc kubenswrapper[4872]: I0126 09:10:14.123512 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5rmt" event={"ID":"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e","Type":"ContainerDied","Data":"cde797e3deb4926f06c3da00432d83090b3025711742fca2c560fc092f42b65b"} Jan 26 09:10:14 crc kubenswrapper[4872]: I0126 09:10:14.135523 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96hdr" event={"ID":"f8525631-7064-4a4e-93ce-a5a5e0f77274","Type":"ContainerStarted","Data":"8587fe7eda98dad781733ca28cb82dca4693de7b85573e7301ff332d7758080a"} Jan 26 09:10:14 crc kubenswrapper[4872]: I0126 09:10:14.138503 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f257f87d723194980907fa8e5aa1966e5008c0d5b9a8292916a0960c8619bbc2"} Jan 26 09:10:14 crc kubenswrapper[4872]: I0126 09:10:14.139384 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"686bcfd8d3ae6ca6e7f2abdd2e8bce2a17db7980e6ce7a503fd730cea98f723f"} Jan 26 09:10:14 crc kubenswrapper[4872]: I0126 09:10:14.146162 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cwfr" event={"ID":"eedb2ebd-5684-464d-bc70-fba04fb6edbd","Type":"ContainerStarted","Data":"e31b0f5619db5b83f281302bcc70d512b01cfc4d5d9d13dd09f9df615c55707c"} Jan 26 09:10:14 crc kubenswrapper[4872]: I0126 09:10:14.151421 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"226bbdcc2d9ab662e3951eed6cf67e4d3bd07a841fa637f7455f6dded07c7453"} Jan 26 09:10:14 crc kubenswrapper[4872]: I0126 09:10:14.153416 4872 generic.go:334] "Generic (PLEG): container finished" podID="bca70073-3d97-452c-818c-70c8021fe37d" containerID="045d5e51a5bf7e4884807e4719a4ee64a765cbbd3784fb5fe58cf4d979c938ec" exitCode=0 Jan 26 09:10:14 crc kubenswrapper[4872]: I0126 09:10:14.153504 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rng9p" event={"ID":"bca70073-3d97-452c-818c-70c8021fe37d","Type":"ContainerDied","Data":"045d5e51a5bf7e4884807e4719a4ee64a765cbbd3784fb5fe58cf4d979c938ec"} Jan 26 09:10:14 crc kubenswrapper[4872]: I0126 09:10:14.163565 4872 generic.go:334] "Generic (PLEG): container finished" podID="c963ddba-b620-46b4-83af-dfbaa96bbd03" containerID="d3df9b0d0f0530aca8ab3604ceae98306edc4cc2cf06f933d1859963c7525c06" exitCode=0 Jan 26 09:10:14 crc kubenswrapper[4872]: I0126 09:10:14.163711 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p76kx" event={"ID":"c963ddba-b620-46b4-83af-dfbaa96bbd03","Type":"ContainerDied","Data":"d3df9b0d0f0530aca8ab3604ceae98306edc4cc2cf06f933d1859963c7525c06"} Jan 26 09:10:14 crc kubenswrapper[4872]: E0126 09:10:14.168247 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-5xspp" podUID="7b6eb987-454b-4c50-bb39-ede2e1006808" Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.175624 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5rmt" event={"ID":"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e","Type":"ContainerStarted","Data":"1003fe652f84611e9a22781903c4dadfbe5af0832431d989b7c8b9ed86fc2811"} Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.178304 4872 generic.go:334] "Generic (PLEG): container finished" podID="f8525631-7064-4a4e-93ce-a5a5e0f77274" containerID="8587fe7eda98dad781733ca28cb82dca4693de7b85573e7301ff332d7758080a" exitCode=0 Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.178415 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96hdr" event={"ID":"f8525631-7064-4a4e-93ce-a5a5e0f77274","Type":"ContainerDied","Data":"8587fe7eda98dad781733ca28cb82dca4693de7b85573e7301ff332d7758080a"} Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.190320 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"cc5c73f029bf9662a4463f25c4fd2b5a3efbaaba267afd2d19b77194e8d97813"} Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.190374 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8db3fed3825a6a7909e431319238cc2dd222473ebd0b7daefe06bc69757ca51f"} Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.190393 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.191224 4872 generic.go:334] "Generic (PLEG): container finished" podID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" containerID="e31b0f5619db5b83f281302bcc70d512b01cfc4d5d9d13dd09f9df615c55707c" exitCode=0 Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.191292 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cwfr" event={"ID":"eedb2ebd-5684-464d-bc70-fba04fb6edbd","Type":"ContainerDied","Data":"e31b0f5619db5b83f281302bcc70d512b01cfc4d5d9d13dd09f9df615c55707c"} Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.198210 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4a98cf3ddb12cfb91cb70c2e8ba94f84a3c6338dd685eadb0f8a8d5a49f72d2f"} Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.203368 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rng9p" event={"ID":"bca70073-3d97-452c-818c-70c8021fe37d","Type":"ContainerStarted","Data":"65d1b5b2a922cde61bd5ec5e1e05ad8d2b56359b267443baf3462f9b0c84e1ed"} Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.206043 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p76kx" event={"ID":"c963ddba-b620-46b4-83af-dfbaa96bbd03","Type":"ContainerStarted","Data":"3fc9adbf84cdcfe30525d46ca6e8e569b105295db734c5e3ad6d0709dded915b"} Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.210438 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p5rmt" podStartSLOduration=2.743257676 podStartE2EDuration="41.21042415s" podCreationTimestamp="2026-01-26 09:09:34 +0000 UTC" firstStartedPulling="2026-01-26 09:09:36.13781111 +0000 UTC m=+109.446650911" lastFinishedPulling="2026-01-26 09:10:14.604977584 +0000 UTC m=+147.913817385" observedRunningTime="2026-01-26 09:10:15.208712412 +0000 UTC m=+148.517552213" watchObservedRunningTime="2026-01-26 09:10:15.21042415 +0000 UTC m=+148.519263951" Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.348376 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rng9p" podStartSLOduration=3.7904463550000003 podStartE2EDuration="42.348351738s" podCreationTimestamp="2026-01-26 09:09:33 +0000 UTC" firstStartedPulling="2026-01-26 09:09:36.123549138 +0000 UTC m=+109.432388939" lastFinishedPulling="2026-01-26 09:10:14.681454521 +0000 UTC m=+147.990294322" observedRunningTime="2026-01-26 09:10:15.346254791 +0000 UTC m=+148.655094592" watchObservedRunningTime="2026-01-26 09:10:15.348351738 +0000 UTC m=+148.657191539" Jan 26 09:10:15 crc kubenswrapper[4872]: I0126 09:10:15.413220 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p76kx" podStartSLOduration=3.820241111 podStartE2EDuration="42.413198364s" podCreationTimestamp="2026-01-26 09:09:33 +0000 UTC" firstStartedPulling="2026-01-26 09:09:36.185462618 +0000 UTC m=+109.494302419" lastFinishedPulling="2026-01-26 09:10:14.778419871 +0000 UTC m=+148.087259672" observedRunningTime="2026-01-26 09:10:15.382752456 +0000 UTC m=+148.691592257" watchObservedRunningTime="2026-01-26 09:10:15.413198364 +0000 UTC m=+148.722038165" Jan 26 09:10:16 crc kubenswrapper[4872]: I0126 09:10:16.215753 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cwfr" event={"ID":"eedb2ebd-5684-464d-bc70-fba04fb6edbd","Type":"ContainerStarted","Data":"eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c"} Jan 26 09:10:16 crc kubenswrapper[4872]: I0126 09:10:16.218334 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96hdr" event={"ID":"f8525631-7064-4a4e-93ce-a5a5e0f77274","Type":"ContainerStarted","Data":"4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b"} Jan 26 09:10:16 crc kubenswrapper[4872]: I0126 09:10:16.236075 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2cwfr" podStartSLOduration=2.848021084 podStartE2EDuration="40.236045388s" podCreationTimestamp="2026-01-26 09:09:36 +0000 UTC" firstStartedPulling="2026-01-26 09:09:38.628530418 +0000 UTC m=+111.937370219" lastFinishedPulling="2026-01-26 09:10:16.016554722 +0000 UTC m=+149.325394523" observedRunningTime="2026-01-26 09:10:16.235492453 +0000 UTC m=+149.544332254" watchObservedRunningTime="2026-01-26 09:10:16.236045388 +0000 UTC m=+149.544885189" Jan 26 09:10:16 crc kubenswrapper[4872]: I0126 09:10:16.261007 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-96hdr" podStartSLOduration=1.914379459 podStartE2EDuration="39.260981815s" podCreationTimestamp="2026-01-26 09:09:37 +0000 UTC" firstStartedPulling="2026-01-26 09:09:38.541306213 +0000 UTC m=+111.850146014" lastFinishedPulling="2026-01-26 09:10:15.887908569 +0000 UTC m=+149.196748370" observedRunningTime="2026-01-26 09:10:16.254597968 +0000 UTC m=+149.563437769" watchObservedRunningTime="2026-01-26 09:10:16.260981815 +0000 UTC m=+149.569821616" Jan 26 09:10:17 crc kubenswrapper[4872]: I0126 09:10:17.084924 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:10:17 crc kubenswrapper[4872]: I0126 09:10:17.085544 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:10:17 crc kubenswrapper[4872]: I0126 09:10:17.505216 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:10:17 crc kubenswrapper[4872]: I0126 09:10:17.506151 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:10:18 crc kubenswrapper[4872]: I0126 09:10:18.150106 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2cwfr" podUID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" containerName="registry-server" probeResult="failure" output=< Jan 26 09:10:18 crc kubenswrapper[4872]: timeout: failed to connect service ":50051" within 1s Jan 26 09:10:18 crc kubenswrapper[4872]: > Jan 26 09:10:18 crc kubenswrapper[4872]: I0126 09:10:18.550438 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-96hdr" podUID="f8525631-7064-4a4e-93ce-a5a5e0f77274" containerName="registry-server" probeResult="failure" output=< Jan 26 09:10:18 crc kubenswrapper[4872]: timeout: failed to connect service ":50051" within 1s Jan 26 09:10:18 crc kubenswrapper[4872]: > Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.726597 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 26 09:10:19 crc kubenswrapper[4872]: E0126 09:10:19.727256 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f3f22d-7f02-4cec-951a-2b05bd144293" containerName="pruner" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.727270 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f3f22d-7f02-4cec-951a-2b05bd144293" containerName="pruner" Jan 26 09:10:19 crc kubenswrapper[4872]: E0126 09:10:19.727322 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bce1b95-50f4-4d87-b7d4-6cae890bfef2" containerName="pruner" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.727328 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bce1b95-50f4-4d87-b7d4-6cae890bfef2" containerName="pruner" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.727442 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="17f3f22d-7f02-4cec-951a-2b05bd144293" containerName="pruner" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.727454 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bce1b95-50f4-4d87-b7d4-6cae890bfef2" containerName="pruner" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.727923 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.730200 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.730485 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.744735 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.842940 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15719b4a-1114-4008-b12d-781975e94b38-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"15719b4a-1114-4008-b12d-781975e94b38\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.843395 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/15719b4a-1114-4008-b12d-781975e94b38-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"15719b4a-1114-4008-b12d-781975e94b38\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.945015 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15719b4a-1114-4008-b12d-781975e94b38-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"15719b4a-1114-4008-b12d-781975e94b38\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.945601 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/15719b4a-1114-4008-b12d-781975e94b38-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"15719b4a-1114-4008-b12d-781975e94b38\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.945742 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/15719b4a-1114-4008-b12d-781975e94b38-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"15719b4a-1114-4008-b12d-781975e94b38\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 09:10:19 crc kubenswrapper[4872]: I0126 09:10:19.972241 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15719b4a-1114-4008-b12d-781975e94b38-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"15719b4a-1114-4008-b12d-781975e94b38\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 09:10:20 crc kubenswrapper[4872]: I0126 09:10:20.057395 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 09:10:20 crc kubenswrapper[4872]: I0126 09:10:20.486098 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Jan 26 09:10:20 crc kubenswrapper[4872]: W0126 09:10:20.491294 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod15719b4a_1114_4008_b12d_781975e94b38.slice/crio-581b52bc995c56b4ebd0aefedbd7852ade998e2a617ae565677846bbdcf93e09 WatchSource:0}: Error finding container 581b52bc995c56b4ebd0aefedbd7852ade998e2a617ae565677846bbdcf93e09: Status 404 returned error can't find the container with id 581b52bc995c56b4ebd0aefedbd7852ade998e2a617ae565677846bbdcf93e09 Jan 26 09:10:21 crc kubenswrapper[4872]: I0126 09:10:21.250762 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"15719b4a-1114-4008-b12d-781975e94b38","Type":"ContainerStarted","Data":"581b52bc995c56b4ebd0aefedbd7852ade998e2a617ae565677846bbdcf93e09"} Jan 26 09:10:23 crc kubenswrapper[4872]: I0126 09:10:23.268354 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"15719b4a-1114-4008-b12d-781975e94b38","Type":"ContainerStarted","Data":"fc15ceb1aa389026df2fe9dcd74a2496416a33f67a31934923febecc400c2a3a"} Jan 26 09:10:23 crc kubenswrapper[4872]: I0126 09:10:23.292275 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=4.29219717 podStartE2EDuration="4.29219717s" podCreationTimestamp="2026-01-26 09:10:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:10:23.291308215 +0000 UTC m=+156.600148016" watchObservedRunningTime="2026-01-26 09:10:23.29219717 +0000 UTC m=+156.601037011" Jan 26 09:10:24 crc kubenswrapper[4872]: I0126 09:10:24.105549 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:10:24 crc kubenswrapper[4872]: I0126 09:10:24.105628 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:10:24 crc kubenswrapper[4872]: I0126 09:10:24.159784 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:10:24 crc kubenswrapper[4872]: I0126 09:10:24.275920 4872 generic.go:334] "Generic (PLEG): container finished" podID="15719b4a-1114-4008-b12d-781975e94b38" containerID="fc15ceb1aa389026df2fe9dcd74a2496416a33f67a31934923febecc400c2a3a" exitCode=0 Jan 26 09:10:24 crc kubenswrapper[4872]: I0126 09:10:24.276035 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"15719b4a-1114-4008-b12d-781975e94b38","Type":"ContainerDied","Data":"fc15ceb1aa389026df2fe9dcd74a2496416a33f67a31934923febecc400c2a3a"} Jan 26 09:10:24 crc kubenswrapper[4872]: I0126 09:10:24.341278 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:10:24 crc kubenswrapper[4872]: I0126 09:10:24.376642 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:10:24 crc kubenswrapper[4872]: I0126 09:10:24.376711 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:10:24 crc kubenswrapper[4872]: I0126 09:10:24.423174 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:10:24 crc kubenswrapper[4872]: I0126 09:10:24.530204 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:10:24 crc kubenswrapper[4872]: I0126 09:10:24.530290 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:10:24 crc kubenswrapper[4872]: I0126 09:10:24.575882 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.333210 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.335811 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.525273 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.526045 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.535309 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.622604 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.638680 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/717a9b10-82c6-4474-8715-0b34ff4e391e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"717a9b10-82c6-4474-8715-0b34ff4e391e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.638743 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/717a9b10-82c6-4474-8715-0b34ff4e391e-kube-api-access\") pod \"installer-9-crc\" (UID: \"717a9b10-82c6-4474-8715-0b34ff4e391e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.638831 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/717a9b10-82c6-4474-8715-0b34ff4e391e-var-lock\") pod \"installer-9-crc\" (UID: \"717a9b10-82c6-4474-8715-0b34ff4e391e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.739683 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15719b4a-1114-4008-b12d-781975e94b38-kube-api-access\") pod \"15719b4a-1114-4008-b12d-781975e94b38\" (UID: \"15719b4a-1114-4008-b12d-781975e94b38\") " Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.739830 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/15719b4a-1114-4008-b12d-781975e94b38-kubelet-dir\") pod \"15719b4a-1114-4008-b12d-781975e94b38\" (UID: \"15719b4a-1114-4008-b12d-781975e94b38\") " Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.739981 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15719b4a-1114-4008-b12d-781975e94b38-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "15719b4a-1114-4008-b12d-781975e94b38" (UID: "15719b4a-1114-4008-b12d-781975e94b38"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.740057 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/717a9b10-82c6-4474-8715-0b34ff4e391e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"717a9b10-82c6-4474-8715-0b34ff4e391e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.740082 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/717a9b10-82c6-4474-8715-0b34ff4e391e-kube-api-access\") pod \"installer-9-crc\" (UID: \"717a9b10-82c6-4474-8715-0b34ff4e391e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.740134 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/717a9b10-82c6-4474-8715-0b34ff4e391e-var-lock\") pod \"installer-9-crc\" (UID: \"717a9b10-82c6-4474-8715-0b34ff4e391e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.740140 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/717a9b10-82c6-4474-8715-0b34ff4e391e-kubelet-dir\") pod \"installer-9-crc\" (UID: \"717a9b10-82c6-4474-8715-0b34ff4e391e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.740214 4872 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/15719b4a-1114-4008-b12d-781975e94b38-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.740237 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/717a9b10-82c6-4474-8715-0b34ff4e391e-var-lock\") pod \"installer-9-crc\" (UID: \"717a9b10-82c6-4474-8715-0b34ff4e391e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.752111 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15719b4a-1114-4008-b12d-781975e94b38-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "15719b4a-1114-4008-b12d-781975e94b38" (UID: "15719b4a-1114-4008-b12d-781975e94b38"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.757698 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/717a9b10-82c6-4474-8715-0b34ff4e391e-kube-api-access\") pod \"installer-9-crc\" (UID: \"717a9b10-82c6-4474-8715-0b34ff4e391e\") " pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.841700 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/15719b4a-1114-4008-b12d-781975e94b38-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:25 crc kubenswrapper[4872]: I0126 09:10:25.866350 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:10:26 crc kubenswrapper[4872]: I0126 09:10:26.119842 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Jan 26 09:10:26 crc kubenswrapper[4872]: I0126 09:10:26.191656 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p5rmt"] Jan 26 09:10:26 crc kubenswrapper[4872]: I0126 09:10:26.295182 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"717a9b10-82c6-4474-8715-0b34ff4e391e","Type":"ContainerStarted","Data":"9ed1f368a28d82c5b6236341331dd0f01711b1eb8cdbe4dea14624463b933e96"} Jan 26 09:10:26 crc kubenswrapper[4872]: I0126 09:10:26.297478 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"15719b4a-1114-4008-b12d-781975e94b38","Type":"ContainerDied","Data":"581b52bc995c56b4ebd0aefedbd7852ade998e2a617ae565677846bbdcf93e09"} Jan 26 09:10:26 crc kubenswrapper[4872]: I0126 09:10:26.297541 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="581b52bc995c56b4ebd0aefedbd7852ade998e2a617ae565677846bbdcf93e09" Jan 26 09:10:26 crc kubenswrapper[4872]: I0126 09:10:26.297875 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Jan 26 09:10:26 crc kubenswrapper[4872]: I0126 09:10:26.792208 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rng9p"] Jan 26 09:10:27 crc kubenswrapper[4872]: I0126 09:10:27.272406 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:10:27 crc kubenswrapper[4872]: I0126 09:10:27.305171 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rng9p" podUID="bca70073-3d97-452c-818c-70c8021fe37d" containerName="registry-server" containerID="cri-o://65d1b5b2a922cde61bd5ec5e1e05ad8d2b56359b267443baf3462f9b0c84e1ed" gracePeriod=2 Jan 26 09:10:27 crc kubenswrapper[4872]: I0126 09:10:27.306149 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"717a9b10-82c6-4474-8715-0b34ff4e391e","Type":"ContainerStarted","Data":"01acc704dd63659f2b630abc8c9febe9217b337227f35452b306fbddd087f4ed"} Jan 26 09:10:27 crc kubenswrapper[4872]: I0126 09:10:27.306297 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p5rmt" podUID="90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" containerName="registry-server" containerID="cri-o://1003fe652f84611e9a22781903c4dadfbe5af0832431d989b7c8b9ed86fc2811" gracePeriod=2 Jan 26 09:10:27 crc kubenswrapper[4872]: I0126 09:10:27.323616 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:10:27 crc kubenswrapper[4872]: I0126 09:10:27.331171 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.331152181 podStartE2EDuration="2.331152181s" podCreationTimestamp="2026-01-26 09:10:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:10:27.326592196 +0000 UTC m=+160.635431997" watchObservedRunningTime="2026-01-26 09:10:27.331152181 +0000 UTC m=+160.639991982" Jan 26 09:10:27 crc kubenswrapper[4872]: I0126 09:10:27.549603 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:10:27 crc kubenswrapper[4872]: I0126 09:10:27.597347 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:10:28 crc kubenswrapper[4872]: I0126 09:10:28.316354 4872 generic.go:334] "Generic (PLEG): container finished" podID="90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" containerID="1003fe652f84611e9a22781903c4dadfbe5af0832431d989b7c8b9ed86fc2811" exitCode=0 Jan 26 09:10:28 crc kubenswrapper[4872]: I0126 09:10:28.316491 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5rmt" event={"ID":"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e","Type":"ContainerDied","Data":"1003fe652f84611e9a22781903c4dadfbe5af0832431d989b7c8b9ed86fc2811"} Jan 26 09:10:28 crc kubenswrapper[4872]: I0126 09:10:28.319988 4872 generic.go:334] "Generic (PLEG): container finished" podID="bca70073-3d97-452c-818c-70c8021fe37d" containerID="65d1b5b2a922cde61bd5ec5e1e05ad8d2b56359b267443baf3462f9b0c84e1ed" exitCode=0 Jan 26 09:10:28 crc kubenswrapper[4872]: I0126 09:10:28.320122 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rng9p" event={"ID":"bca70073-3d97-452c-818c-70c8021fe37d","Type":"ContainerDied","Data":"65d1b5b2a922cde61bd5ec5e1e05ad8d2b56359b267443baf3462f9b0c84e1ed"} Jan 26 09:10:29 crc kubenswrapper[4872]: I0126 09:10:29.103286 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:10:29 crc kubenswrapper[4872]: I0126 09:10:29.106216 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 26 09:10:29 crc kubenswrapper[4872]: I0126 09:10:29.123728 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f34a04d-444c-4640-9b51-51850a6317e6-metrics-certs\") pod \"network-metrics-daemon-krd2b\" (UID: \"8f34a04d-444c-4640-9b51-51850a6317e6\") " pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:10:29 crc kubenswrapper[4872]: I0126 09:10:29.429891 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 26 09:10:29 crc kubenswrapper[4872]: I0126 09:10:29.431205 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krd2b" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.037037 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.109194 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.178525 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-krd2b"] Jan 26 09:10:31 crc kubenswrapper[4872]: W0126 09:10:31.192077 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f34a04d_444c_4640_9b51_51850a6317e6.slice/crio-49a3eb831d23234b99dacede60c98143e6ba26ff94eb7ea9532d9e69c132e2c1 WatchSource:0}: Error finding container 49a3eb831d23234b99dacede60c98143e6ba26ff94eb7ea9532d9e69c132e2c1: Status 404 returned error can't find the container with id 49a3eb831d23234b99dacede60c98143e6ba26ff94eb7ea9532d9e69c132e2c1 Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.233202 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnswd\" (UniqueName: \"kubernetes.io/projected/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-kube-api-access-fnswd\") pod \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\" (UID: \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\") " Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.233283 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bca70073-3d97-452c-818c-70c8021fe37d-catalog-content\") pod \"bca70073-3d97-452c-818c-70c8021fe37d\" (UID: \"bca70073-3d97-452c-818c-70c8021fe37d\") " Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.233318 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bca70073-3d97-452c-818c-70c8021fe37d-utilities\") pod \"bca70073-3d97-452c-818c-70c8021fe37d\" (UID: \"bca70073-3d97-452c-818c-70c8021fe37d\") " Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.233419 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-catalog-content\") pod \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\" (UID: \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\") " Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.233487 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtx4m\" (UniqueName: \"kubernetes.io/projected/bca70073-3d97-452c-818c-70c8021fe37d-kube-api-access-mtx4m\") pod \"bca70073-3d97-452c-818c-70c8021fe37d\" (UID: \"bca70073-3d97-452c-818c-70c8021fe37d\") " Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.234625 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bca70073-3d97-452c-818c-70c8021fe37d-utilities" (OuterVolumeSpecName: "utilities") pod "bca70073-3d97-452c-818c-70c8021fe37d" (UID: "bca70073-3d97-452c-818c-70c8021fe37d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.234897 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-utilities" (OuterVolumeSpecName: "utilities") pod "90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" (UID: "90bc1e8f-1522-4f57-afe3-e3745b7d6f3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.240257 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-kube-api-access-fnswd" (OuterVolumeSpecName: "kube-api-access-fnswd") pod "90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" (UID: "90bc1e8f-1522-4f57-afe3-e3745b7d6f3e"). InnerVolumeSpecName "kube-api-access-fnswd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.240380 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bca70073-3d97-452c-818c-70c8021fe37d-kube-api-access-mtx4m" (OuterVolumeSpecName: "kube-api-access-mtx4m") pod "bca70073-3d97-452c-818c-70c8021fe37d" (UID: "bca70073-3d97-452c-818c-70c8021fe37d"). InnerVolumeSpecName "kube-api-access-mtx4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.243856 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-utilities\") pod \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\" (UID: \"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e\") " Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.244389 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnswd\" (UniqueName: \"kubernetes.io/projected/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-kube-api-access-fnswd\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.244417 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bca70073-3d97-452c-818c-70c8021fe37d-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.244434 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtx4m\" (UniqueName: \"kubernetes.io/projected/bca70073-3d97-452c-818c-70c8021fe37d-kube-api-access-mtx4m\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.244446 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.283506 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bca70073-3d97-452c-818c-70c8021fe37d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bca70073-3d97-452c-818c-70c8021fe37d" (UID: "bca70073-3d97-452c-818c-70c8021fe37d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.284853 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" (UID: "90bc1e8f-1522-4f57-afe3-e3745b7d6f3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.345930 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bca70073-3d97-452c-818c-70c8021fe37d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.345991 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.349603 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p5rmt" event={"ID":"90bc1e8f-1522-4f57-afe3-e3745b7d6f3e","Type":"ContainerDied","Data":"dbc5d37843ef6b28528a3abc21ad4cf93d711a95a81240dd10563bee59d67ba0"} Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.349697 4872 scope.go:117] "RemoveContainer" containerID="1003fe652f84611e9a22781903c4dadfbe5af0832431d989b7c8b9ed86fc2811" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.349649 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p5rmt" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.351107 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-krd2b" event={"ID":"8f34a04d-444c-4640-9b51-51850a6317e6","Type":"ContainerStarted","Data":"49a3eb831d23234b99dacede60c98143e6ba26ff94eb7ea9532d9e69c132e2c1"} Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.360236 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rng9p" event={"ID":"bca70073-3d97-452c-818c-70c8021fe37d","Type":"ContainerDied","Data":"3888844e8ed08ca807359e7b74d94ba3d4424b7f9b29d1e659f162e20c744cd5"} Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.360332 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rng9p" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.369462 4872 scope.go:117] "RemoveContainer" containerID="cde797e3deb4926f06c3da00432d83090b3025711742fca2c560fc092f42b65b" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.402578 4872 scope.go:117] "RemoveContainer" containerID="63de97ba6862e1a08f8604dd1fbf0c30372ae1b3141813698bbb87ac658c7630" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.414013 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rng9p"] Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.418449 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rng9p"] Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.429902 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p5rmt"] Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.434695 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p5rmt"] Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.436526 4872 scope.go:117] "RemoveContainer" containerID="65d1b5b2a922cde61bd5ec5e1e05ad8d2b56359b267443baf3462f9b0c84e1ed" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.450565 4872 scope.go:117] "RemoveContainer" containerID="045d5e51a5bf7e4884807e4719a4ee64a765cbbd3784fb5fe58cf4d979c938ec" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.480012 4872 scope.go:117] "RemoveContainer" containerID="7381fa6fe6e23850bb1f8bce9429b5f6588df3f83479b198b1f849752a859a7a" Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.597292 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-96hdr"] Jan 26 09:10:31 crc kubenswrapper[4872]: I0126 09:10:31.597938 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-96hdr" podUID="f8525631-7064-4a4e-93ce-a5a5e0f77274" containerName="registry-server" containerID="cri-o://4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b" gracePeriod=2 Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.041705 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.055736 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8525631-7064-4a4e-93ce-a5a5e0f77274-catalog-content\") pod \"f8525631-7064-4a4e-93ce-a5a5e0f77274\" (UID: \"f8525631-7064-4a4e-93ce-a5a5e0f77274\") " Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.055838 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8525631-7064-4a4e-93ce-a5a5e0f77274-utilities\") pod \"f8525631-7064-4a4e-93ce-a5a5e0f77274\" (UID: \"f8525631-7064-4a4e-93ce-a5a5e0f77274\") " Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.055928 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcn9t\" (UniqueName: \"kubernetes.io/projected/f8525631-7064-4a4e-93ce-a5a5e0f77274-kube-api-access-xcn9t\") pod \"f8525631-7064-4a4e-93ce-a5a5e0f77274\" (UID: \"f8525631-7064-4a4e-93ce-a5a5e0f77274\") " Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.057998 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8525631-7064-4a4e-93ce-a5a5e0f77274-utilities" (OuterVolumeSpecName: "utilities") pod "f8525631-7064-4a4e-93ce-a5a5e0f77274" (UID: "f8525631-7064-4a4e-93ce-a5a5e0f77274"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.063627 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8525631-7064-4a4e-93ce-a5a5e0f77274-kube-api-access-xcn9t" (OuterVolumeSpecName: "kube-api-access-xcn9t") pod "f8525631-7064-4a4e-93ce-a5a5e0f77274" (UID: "f8525631-7064-4a4e-93ce-a5a5e0f77274"). InnerVolumeSpecName "kube-api-access-xcn9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.158016 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8525631-7064-4a4e-93ce-a5a5e0f77274-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.158114 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcn9t\" (UniqueName: \"kubernetes.io/projected/f8525631-7064-4a4e-93ce-a5a5e0f77274-kube-api-access-xcn9t\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.203393 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8525631-7064-4a4e-93ce-a5a5e0f77274-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8525631-7064-4a4e-93ce-a5a5e0f77274" (UID: "f8525631-7064-4a4e-93ce-a5a5e0f77274"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.259544 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8525631-7064-4a4e-93ce-a5a5e0f77274-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.390821 4872 generic.go:334] "Generic (PLEG): container finished" podID="f8525631-7064-4a4e-93ce-a5a5e0f77274" containerID="4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b" exitCode=0 Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.390912 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96hdr" event={"ID":"f8525631-7064-4a4e-93ce-a5a5e0f77274","Type":"ContainerDied","Data":"4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b"} Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.390991 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-96hdr" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.391014 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-96hdr" event={"ID":"f8525631-7064-4a4e-93ce-a5a5e0f77274","Type":"ContainerDied","Data":"befe57ef566e35a3767e5693dddbe843d8b0ea10ea0e41ba7a2178633e7f3e40"} Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.391041 4872 scope.go:117] "RemoveContainer" containerID="4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.394340 4872 generic.go:334] "Generic (PLEG): container finished" podID="7b6eb987-454b-4c50-bb39-ede2e1006808" containerID="759ad08f8a02bf6c6af28abcbb2033a94c818e363d07b748bf476343b8ee23c6" exitCode=0 Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.394426 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xspp" event={"ID":"7b6eb987-454b-4c50-bb39-ede2e1006808","Type":"ContainerDied","Data":"759ad08f8a02bf6c6af28abcbb2033a94c818e363d07b748bf476343b8ee23c6"} Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.397588 4872 generic.go:334] "Generic (PLEG): container finished" podID="3132ab11-2629-4089-b4e5-da75a4d83fd9" containerID="e5670ba6e88f896aee3a27ccec89990a2ab2556dce15db873a5622043e12c7fc" exitCode=0 Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.397648 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84qrq" event={"ID":"3132ab11-2629-4089-b4e5-da75a4d83fd9","Type":"ContainerDied","Data":"e5670ba6e88f896aee3a27ccec89990a2ab2556dce15db873a5622043e12c7fc"} Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.401820 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-krd2b" event={"ID":"8f34a04d-444c-4640-9b51-51850a6317e6","Type":"ContainerStarted","Data":"3043ac0e1883b15464ba1f59a05cee3209dad74834df57897e2e2593576bf0a5"} Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.401871 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-krd2b" event={"ID":"8f34a04d-444c-4640-9b51-51850a6317e6","Type":"ContainerStarted","Data":"0cff03370cb491cf024dc15ed79c4282e5fcd07745f970293bd099c1d4739621"} Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.408277 4872 generic.go:334] "Generic (PLEG): container finished" podID="2d39d8ba-e193-4c19-a535-589ffd091413" containerID="447d098c44565af51e34eb927cd0e41b6feeb679ca51037c6bb9b95ccde8f8bd" exitCode=0 Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.408375 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qfx7" event={"ID":"2d39d8ba-e193-4c19-a535-589ffd091413","Type":"ContainerDied","Data":"447d098c44565af51e34eb927cd0e41b6feeb679ca51037c6bb9b95ccde8f8bd"} Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.427962 4872 scope.go:117] "RemoveContainer" containerID="8587fe7eda98dad781733ca28cb82dca4693de7b85573e7301ff332d7758080a" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.484562 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-krd2b" podStartSLOduration=146.484537147 podStartE2EDuration="2m26.484537147s" podCreationTimestamp="2026-01-26 09:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:10:32.483669072 +0000 UTC m=+165.792508883" watchObservedRunningTime="2026-01-26 09:10:32.484537147 +0000 UTC m=+165.793376968" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.485125 4872 scope.go:117] "RemoveContainer" containerID="713be9d4ef286298e87284c52944fc06c70c5bf43723456d532461662d378ea6" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.507395 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-96hdr"] Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.512823 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-96hdr"] Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.515641 4872 scope.go:117] "RemoveContainer" containerID="4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b" Jan 26 09:10:32 crc kubenswrapper[4872]: E0126 09:10:32.519361 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b\": container with ID starting with 4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b not found: ID does not exist" containerID="4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.519441 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b"} err="failed to get container status \"4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b\": rpc error: code = NotFound desc = could not find container \"4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b\": container with ID starting with 4242e04d7d38bee04489a2a1dd0a1b8a26defc6f40c6ab368237a5bb3e60d73b not found: ID does not exist" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.519539 4872 scope.go:117] "RemoveContainer" containerID="8587fe7eda98dad781733ca28cb82dca4693de7b85573e7301ff332d7758080a" Jan 26 09:10:32 crc kubenswrapper[4872]: E0126 09:10:32.520135 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8587fe7eda98dad781733ca28cb82dca4693de7b85573e7301ff332d7758080a\": container with ID starting with 8587fe7eda98dad781733ca28cb82dca4693de7b85573e7301ff332d7758080a not found: ID does not exist" containerID="8587fe7eda98dad781733ca28cb82dca4693de7b85573e7301ff332d7758080a" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.520178 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8587fe7eda98dad781733ca28cb82dca4693de7b85573e7301ff332d7758080a"} err="failed to get container status \"8587fe7eda98dad781733ca28cb82dca4693de7b85573e7301ff332d7758080a\": rpc error: code = NotFound desc = could not find container \"8587fe7eda98dad781733ca28cb82dca4693de7b85573e7301ff332d7758080a\": container with ID starting with 8587fe7eda98dad781733ca28cb82dca4693de7b85573e7301ff332d7758080a not found: ID does not exist" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.520218 4872 scope.go:117] "RemoveContainer" containerID="713be9d4ef286298e87284c52944fc06c70c5bf43723456d532461662d378ea6" Jan 26 09:10:32 crc kubenswrapper[4872]: E0126 09:10:32.520698 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"713be9d4ef286298e87284c52944fc06c70c5bf43723456d532461662d378ea6\": container with ID starting with 713be9d4ef286298e87284c52944fc06c70c5bf43723456d532461662d378ea6 not found: ID does not exist" containerID="713be9d4ef286298e87284c52944fc06c70c5bf43723456d532461662d378ea6" Jan 26 09:10:32 crc kubenswrapper[4872]: I0126 09:10:32.520726 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"713be9d4ef286298e87284c52944fc06c70c5bf43723456d532461662d378ea6"} err="failed to get container status \"713be9d4ef286298e87284c52944fc06c70c5bf43723456d532461662d378ea6\": rpc error: code = NotFound desc = could not find container \"713be9d4ef286298e87284c52944fc06c70c5bf43723456d532461662d378ea6\": container with ID starting with 713be9d4ef286298e87284c52944fc06c70c5bf43723456d532461662d378ea6 not found: ID does not exist" Jan 26 09:10:33 crc kubenswrapper[4872]: I0126 09:10:33.192136 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" path="/var/lib/kubelet/pods/90bc1e8f-1522-4f57-afe3-e3745b7d6f3e/volumes" Jan 26 09:10:33 crc kubenswrapper[4872]: I0126 09:10:33.193211 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bca70073-3d97-452c-818c-70c8021fe37d" path="/var/lib/kubelet/pods/bca70073-3d97-452c-818c-70c8021fe37d/volumes" Jan 26 09:10:33 crc kubenswrapper[4872]: I0126 09:10:33.194310 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8525631-7064-4a4e-93ce-a5a5e0f77274" path="/var/lib/kubelet/pods/f8525631-7064-4a4e-93ce-a5a5e0f77274/volumes" Jan 26 09:10:33 crc kubenswrapper[4872]: I0126 09:10:33.421595 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xspp" event={"ID":"7b6eb987-454b-4c50-bb39-ede2e1006808","Type":"ContainerStarted","Data":"f2d39ac93259445624371a26a04d3f7afed05a1b2b4838aceac1951edf35426a"} Jan 26 09:10:33 crc kubenswrapper[4872]: I0126 09:10:33.423918 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84qrq" event={"ID":"3132ab11-2629-4089-b4e5-da75a4d83fd9","Type":"ContainerStarted","Data":"f5b4fa1964841217ba62256094305892512dda59254f7b3046755749a7addadf"} Jan 26 09:10:33 crc kubenswrapper[4872]: I0126 09:10:33.427533 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qfx7" event={"ID":"2d39d8ba-e193-4c19-a535-589ffd091413","Type":"ContainerStarted","Data":"2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332"} Jan 26 09:10:33 crc kubenswrapper[4872]: I0126 09:10:33.448365 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5xspp" podStartSLOduration=3.64398672 podStartE2EDuration="1m0.448340841s" podCreationTimestamp="2026-01-26 09:09:33 +0000 UTC" firstStartedPulling="2026-01-26 09:09:36.142098548 +0000 UTC m=+109.450938339" lastFinishedPulling="2026-01-26 09:10:32.946452649 +0000 UTC m=+166.255292460" observedRunningTime="2026-01-26 09:10:33.44355816 +0000 UTC m=+166.752397961" watchObservedRunningTime="2026-01-26 09:10:33.448340841 +0000 UTC m=+166.757180642" Jan 26 09:10:33 crc kubenswrapper[4872]: I0126 09:10:33.489743 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5qfx7" podStartSLOduration=4.223119176 podStartE2EDuration="58.489724542s" podCreationTimestamp="2026-01-26 09:09:35 +0000 UTC" firstStartedPulling="2026-01-26 09:09:38.628913789 +0000 UTC m=+111.937753580" lastFinishedPulling="2026-01-26 09:10:32.895519145 +0000 UTC m=+166.204358946" observedRunningTime="2026-01-26 09:10:33.488370214 +0000 UTC m=+166.797210025" watchObservedRunningTime="2026-01-26 09:10:33.489724542 +0000 UTC m=+166.798564343" Jan 26 09:10:33 crc kubenswrapper[4872]: I0126 09:10:33.491201 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-84qrq" podStartSLOduration=3.087810501 podStartE2EDuration="58.491192952s" podCreationTimestamp="2026-01-26 09:09:35 +0000 UTC" firstStartedPulling="2026-01-26 09:09:37.398244949 +0000 UTC m=+110.707084750" lastFinishedPulling="2026-01-26 09:10:32.8016274 +0000 UTC m=+166.110467201" observedRunningTime="2026-01-26 09:10:33.46789978 +0000 UTC m=+166.776739581" watchObservedRunningTime="2026-01-26 09:10:33.491192952 +0000 UTC m=+166.800032753" Jan 26 09:10:34 crc kubenswrapper[4872]: I0126 09:10:34.004762 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:10:34 crc kubenswrapper[4872]: I0126 09:10:34.004845 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:10:35 crc kubenswrapper[4872]: I0126 09:10:35.049092 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-5xspp" podUID="7b6eb987-454b-4c50-bb39-ede2e1006808" containerName="registry-server" probeResult="failure" output=< Jan 26 09:10:35 crc kubenswrapper[4872]: timeout: failed to connect service ":50051" within 1s Jan 26 09:10:35 crc kubenswrapper[4872]: > Jan 26 09:10:35 crc kubenswrapper[4872]: I0126 09:10:35.953857 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:10:35 crc kubenswrapper[4872]: I0126 09:10:35.953955 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:10:35 crc kubenswrapper[4872]: I0126 09:10:35.999724 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:10:36 crc kubenswrapper[4872]: I0126 09:10:36.196721 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:10:36 crc kubenswrapper[4872]: I0126 09:10:36.196831 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:10:36 crc kubenswrapper[4872]: I0126 09:10:36.237043 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:10:40 crc kubenswrapper[4872]: I0126 09:10:40.227988 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:10:40 crc kubenswrapper[4872]: I0126 09:10:40.228426 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:10:44 crc kubenswrapper[4872]: I0126 09:10:44.057433 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:10:44 crc kubenswrapper[4872]: I0126 09:10:44.106378 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:10:46 crc kubenswrapper[4872]: I0126 09:10:46.002816 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:10:46 crc kubenswrapper[4872]: I0126 09:10:46.239234 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:10:48 crc kubenswrapper[4872]: I0126 09:10:48.489037 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qfx7"] Jan 26 09:10:48 crc kubenswrapper[4872]: I0126 09:10:48.489852 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5qfx7" podUID="2d39d8ba-e193-4c19-a535-589ffd091413" containerName="registry-server" containerID="cri-o://2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332" gracePeriod=2 Jan 26 09:10:48 crc kubenswrapper[4872]: I0126 09:10:48.901627 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.023428 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sm4rl\" (UniqueName: \"kubernetes.io/projected/2d39d8ba-e193-4c19-a535-589ffd091413-kube-api-access-sm4rl\") pod \"2d39d8ba-e193-4c19-a535-589ffd091413\" (UID: \"2d39d8ba-e193-4c19-a535-589ffd091413\") " Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.023545 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d39d8ba-e193-4c19-a535-589ffd091413-catalog-content\") pod \"2d39d8ba-e193-4c19-a535-589ffd091413\" (UID: \"2d39d8ba-e193-4c19-a535-589ffd091413\") " Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.023614 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d39d8ba-e193-4c19-a535-589ffd091413-utilities\") pod \"2d39d8ba-e193-4c19-a535-589ffd091413\" (UID: \"2d39d8ba-e193-4c19-a535-589ffd091413\") " Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.024719 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d39d8ba-e193-4c19-a535-589ffd091413-utilities" (OuterVolumeSpecName: "utilities") pod "2d39d8ba-e193-4c19-a535-589ffd091413" (UID: "2d39d8ba-e193-4c19-a535-589ffd091413"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.029844 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d39d8ba-e193-4c19-a535-589ffd091413-kube-api-access-sm4rl" (OuterVolumeSpecName: "kube-api-access-sm4rl") pod "2d39d8ba-e193-4c19-a535-589ffd091413" (UID: "2d39d8ba-e193-4c19-a535-589ffd091413"). InnerVolumeSpecName "kube-api-access-sm4rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.047834 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d39d8ba-e193-4c19-a535-589ffd091413-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d39d8ba-e193-4c19-a535-589ffd091413" (UID: "2d39d8ba-e193-4c19-a535-589ffd091413"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.125080 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d39d8ba-e193-4c19-a535-589ffd091413-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.125129 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sm4rl\" (UniqueName: \"kubernetes.io/projected/2d39d8ba-e193-4c19-a535-589ffd091413-kube-api-access-sm4rl\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.125142 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d39d8ba-e193-4c19-a535-589ffd091413-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.550548 4872 generic.go:334] "Generic (PLEG): container finished" podID="2d39d8ba-e193-4c19-a535-589ffd091413" containerID="2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332" exitCode=0 Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.550609 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qfx7" event={"ID":"2d39d8ba-e193-4c19-a535-589ffd091413","Type":"ContainerDied","Data":"2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332"} Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.550650 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5qfx7" event={"ID":"2d39d8ba-e193-4c19-a535-589ffd091413","Type":"ContainerDied","Data":"43b03334a8ba72b9d7d3e396fc432fb4d859a01f1c617018396074aa085148d7"} Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.550677 4872 scope.go:117] "RemoveContainer" containerID="2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.550688 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5qfx7" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.570638 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qfx7"] Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.573167 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5qfx7"] Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.577600 4872 scope.go:117] "RemoveContainer" containerID="447d098c44565af51e34eb927cd0e41b6feeb679ca51037c6bb9b95ccde8f8bd" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.605231 4872 scope.go:117] "RemoveContainer" containerID="0c19873dc62882d3c604e85ef7e900b7d9add7cf21d701109258e9bd690f7ea9" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.624562 4872 scope.go:117] "RemoveContainer" containerID="2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332" Jan 26 09:10:49 crc kubenswrapper[4872]: E0126 09:10:49.624954 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332\": container with ID starting with 2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332 not found: ID does not exist" containerID="2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.624990 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332"} err="failed to get container status \"2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332\": rpc error: code = NotFound desc = could not find container \"2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332\": container with ID starting with 2cf62fea788692667660ca7c70f9249bd7f418406add866057c9f7f1d89e3332 not found: ID does not exist" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.625013 4872 scope.go:117] "RemoveContainer" containerID="447d098c44565af51e34eb927cd0e41b6feeb679ca51037c6bb9b95ccde8f8bd" Jan 26 09:10:49 crc kubenswrapper[4872]: E0126 09:10:49.625318 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"447d098c44565af51e34eb927cd0e41b6feeb679ca51037c6bb9b95ccde8f8bd\": container with ID starting with 447d098c44565af51e34eb927cd0e41b6feeb679ca51037c6bb9b95ccde8f8bd not found: ID does not exist" containerID="447d098c44565af51e34eb927cd0e41b6feeb679ca51037c6bb9b95ccde8f8bd" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.625388 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"447d098c44565af51e34eb927cd0e41b6feeb679ca51037c6bb9b95ccde8f8bd"} err="failed to get container status \"447d098c44565af51e34eb927cd0e41b6feeb679ca51037c6bb9b95ccde8f8bd\": rpc error: code = NotFound desc = could not find container \"447d098c44565af51e34eb927cd0e41b6feeb679ca51037c6bb9b95ccde8f8bd\": container with ID starting with 447d098c44565af51e34eb927cd0e41b6feeb679ca51037c6bb9b95ccde8f8bd not found: ID does not exist" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.625438 4872 scope.go:117] "RemoveContainer" containerID="0c19873dc62882d3c604e85ef7e900b7d9add7cf21d701109258e9bd690f7ea9" Jan 26 09:10:49 crc kubenswrapper[4872]: E0126 09:10:49.626063 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c19873dc62882d3c604e85ef7e900b7d9add7cf21d701109258e9bd690f7ea9\": container with ID starting with 0c19873dc62882d3c604e85ef7e900b7d9add7cf21d701109258e9bd690f7ea9 not found: ID does not exist" containerID="0c19873dc62882d3c604e85ef7e900b7d9add7cf21d701109258e9bd690f7ea9" Jan 26 09:10:49 crc kubenswrapper[4872]: I0126 09:10:49.626100 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c19873dc62882d3c604e85ef7e900b7d9add7cf21d701109258e9bd690f7ea9"} err="failed to get container status \"0c19873dc62882d3c604e85ef7e900b7d9add7cf21d701109258e9bd690f7ea9\": rpc error: code = NotFound desc = could not find container \"0c19873dc62882d3c604e85ef7e900b7d9add7cf21d701109258e9bd690f7ea9\": container with ID starting with 0c19873dc62882d3c604e85ef7e900b7d9add7cf21d701109258e9bd690f7ea9 not found: ID does not exist" Jan 26 09:10:51 crc kubenswrapper[4872]: I0126 09:10:51.190780 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d39d8ba-e193-4c19-a535-589ffd091413" path="/var/lib/kubelet/pods/2d39d8ba-e193-4c19-a535-589ffd091413/volumes" Jan 26 09:10:51 crc kubenswrapper[4872]: I0126 09:10:51.847374 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mk4bd"] Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.009608 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p76kx"] Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.012233 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p76kx" podUID="c963ddba-b620-46b4-83af-dfbaa96bbd03" containerName="registry-server" containerID="cri-o://3fc9adbf84cdcfe30525d46ca6e8e569b105295db734c5e3ad6d0709dded915b" gracePeriod=30 Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.027675 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5xspp"] Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.028426 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5xspp" podUID="7b6eb987-454b-4c50-bb39-ede2e1006808" containerName="registry-server" containerID="cri-o://f2d39ac93259445624371a26a04d3f7afed05a1b2b4838aceac1951edf35426a" gracePeriod=30 Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.057861 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bsfdw"] Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.058143 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" podUID="07b7e8f0-a421-4fbc-9c0d-aa773210edcd" containerName="marketplace-operator" containerID="cri-o://0f074f626e397d4d0e1b5b6e4c1aecb86c1a34078c0ce317ae30c6ebcfba7e34" gracePeriod=30 Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.060652 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-84qrq"] Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.061103 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-84qrq" podUID="3132ab11-2629-4089-b4e5-da75a4d83fd9" containerName="registry-server" containerID="cri-o://f5b4fa1964841217ba62256094305892512dda59254f7b3046755749a7addadf" gracePeriod=30 Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063288 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9ckvb"] Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063540 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15719b4a-1114-4008-b12d-781975e94b38" containerName="pruner" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063557 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="15719b4a-1114-4008-b12d-781975e94b38" containerName="pruner" Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063569 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bca70073-3d97-452c-818c-70c8021fe37d" containerName="registry-server" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063577 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="bca70073-3d97-452c-818c-70c8021fe37d" containerName="registry-server" Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063586 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" containerName="extract-utilities" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063592 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" containerName="extract-utilities" Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063605 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8525631-7064-4a4e-93ce-a5a5e0f77274" containerName="registry-server" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063610 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8525631-7064-4a4e-93ce-a5a5e0f77274" containerName="registry-server" Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063619 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d39d8ba-e193-4c19-a535-589ffd091413" containerName="extract-utilities" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063625 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d39d8ba-e193-4c19-a535-589ffd091413" containerName="extract-utilities" Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063633 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bca70073-3d97-452c-818c-70c8021fe37d" containerName="extract-content" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063638 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="bca70073-3d97-452c-818c-70c8021fe37d" containerName="extract-content" Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063645 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8525631-7064-4a4e-93ce-a5a5e0f77274" containerName="extract-utilities" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063651 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8525631-7064-4a4e-93ce-a5a5e0f77274" containerName="extract-utilities" Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063659 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" containerName="extract-content" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063666 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" containerName="extract-content" Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063673 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d39d8ba-e193-4c19-a535-589ffd091413" containerName="extract-content" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063679 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d39d8ba-e193-4c19-a535-589ffd091413" containerName="extract-content" Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063687 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bca70073-3d97-452c-818c-70c8021fe37d" containerName="extract-utilities" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063693 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="bca70073-3d97-452c-818c-70c8021fe37d" containerName="extract-utilities" Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063702 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" containerName="registry-server" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063709 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" containerName="registry-server" Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063720 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d39d8ba-e193-4c19-a535-589ffd091413" containerName="registry-server" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063725 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d39d8ba-e193-4c19-a535-589ffd091413" containerName="registry-server" Jan 26 09:10:53 crc kubenswrapper[4872]: E0126 09:10:53.063735 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8525631-7064-4a4e-93ce-a5a5e0f77274" containerName="extract-content" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063740 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8525631-7064-4a4e-93ce-a5a5e0f77274" containerName="extract-content" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063846 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="bca70073-3d97-452c-818c-70c8021fe37d" containerName="registry-server" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063856 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="90bc1e8f-1522-4f57-afe3-e3745b7d6f3e" containerName="registry-server" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063865 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d39d8ba-e193-4c19-a535-589ffd091413" containerName="registry-server" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063874 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8525631-7064-4a4e-93ce-a5a5e0f77274" containerName="registry-server" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.063885 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="15719b4a-1114-4008-b12d-781975e94b38" containerName="pruner" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.064353 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.066154 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2cwfr"] Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.066518 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2cwfr" podUID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" containerName="registry-server" containerID="cri-o://eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c" gracePeriod=30 Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.068502 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9ckvb"] Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.184419 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74ttz\" (UniqueName: \"kubernetes.io/projected/579725e4-42fb-49f8-b445-4eed08a373b6-kube-api-access-74ttz\") pod \"marketplace-operator-79b997595-9ckvb\" (UID: \"579725e4-42fb-49f8-b445-4eed08a373b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.184490 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/579725e4-42fb-49f8-b445-4eed08a373b6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9ckvb\" (UID: \"579725e4-42fb-49f8-b445-4eed08a373b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.184549 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/579725e4-42fb-49f8-b445-4eed08a373b6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9ckvb\" (UID: \"579725e4-42fb-49f8-b445-4eed08a373b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.286616 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74ttz\" (UniqueName: \"kubernetes.io/projected/579725e4-42fb-49f8-b445-4eed08a373b6-kube-api-access-74ttz\") pod \"marketplace-operator-79b997595-9ckvb\" (UID: \"579725e4-42fb-49f8-b445-4eed08a373b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.286693 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/579725e4-42fb-49f8-b445-4eed08a373b6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9ckvb\" (UID: \"579725e4-42fb-49f8-b445-4eed08a373b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.286714 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/579725e4-42fb-49f8-b445-4eed08a373b6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9ckvb\" (UID: \"579725e4-42fb-49f8-b445-4eed08a373b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.289316 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/579725e4-42fb-49f8-b445-4eed08a373b6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-9ckvb\" (UID: \"579725e4-42fb-49f8-b445-4eed08a373b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.294965 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/579725e4-42fb-49f8-b445-4eed08a373b6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-9ckvb\" (UID: \"579725e4-42fb-49f8-b445-4eed08a373b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.307193 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74ttz\" (UniqueName: \"kubernetes.io/projected/579725e4-42fb-49f8-b445-4eed08a373b6-kube-api-access-74ttz\") pod \"marketplace-operator-79b997595-9ckvb\" (UID: \"579725e4-42fb-49f8-b445-4eed08a373b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.388420 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.495521 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.590912 4872 generic.go:334] "Generic (PLEG): container finished" podID="7b6eb987-454b-4c50-bb39-ede2e1006808" containerID="f2d39ac93259445624371a26a04d3f7afed05a1b2b4838aceac1951edf35426a" exitCode=0 Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.591014 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xspp" event={"ID":"7b6eb987-454b-4c50-bb39-ede2e1006808","Type":"ContainerDied","Data":"f2d39ac93259445624371a26a04d3f7afed05a1b2b4838aceac1951edf35426a"} Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.599590 4872 generic.go:334] "Generic (PLEG): container finished" podID="3132ab11-2629-4089-b4e5-da75a4d83fd9" containerID="f5b4fa1964841217ba62256094305892512dda59254f7b3046755749a7addadf" exitCode=0 Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.599671 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84qrq" event={"ID":"3132ab11-2629-4089-b4e5-da75a4d83fd9","Type":"ContainerDied","Data":"f5b4fa1964841217ba62256094305892512dda59254f7b3046755749a7addadf"} Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.604744 4872 generic.go:334] "Generic (PLEG): container finished" podID="07b7e8f0-a421-4fbc-9c0d-aa773210edcd" containerID="0f074f626e397d4d0e1b5b6e4c1aecb86c1a34078c0ce317ae30c6ebcfba7e34" exitCode=0 Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.605021 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" event={"ID":"07b7e8f0-a421-4fbc-9c0d-aa773210edcd","Type":"ContainerDied","Data":"0f074f626e397d4d0e1b5b6e4c1aecb86c1a34078c0ce317ae30c6ebcfba7e34"} Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.609657 4872 generic.go:334] "Generic (PLEG): container finished" podID="c963ddba-b620-46b4-83af-dfbaa96bbd03" containerID="3fc9adbf84cdcfe30525d46ca6e8e569b105295db734c5e3ad6d0709dded915b" exitCode=0 Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.609696 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p76kx" event={"ID":"c963ddba-b620-46b4-83af-dfbaa96bbd03","Type":"ContainerDied","Data":"3fc9adbf84cdcfe30525d46ca6e8e569b105295db734c5e3ad6d0709dded915b"} Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.839455 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-9ckvb"] Jan 26 09:10:53 crc kubenswrapper[4872]: W0126 09:10:53.877320 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod579725e4_42fb_49f8_b445_4eed08a373b6.slice/crio-fdc1486d28b0568fa6335022262f526d183b0c91444a169229f3e49742a6c46d WatchSource:0}: Error finding container fdc1486d28b0568fa6335022262f526d183b0c91444a169229f3e49742a6c46d: Status 404 returned error can't find the container with id fdc1486d28b0568fa6335022262f526d183b0c91444a169229f3e49742a6c46d Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.959466 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:10:53 crc kubenswrapper[4872]: I0126 09:10:53.996481 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.009513 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-utilities\") pod \"7b6eb987-454b-4c50-bb39-ede2e1006808\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.009584 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-catalog-content\") pod \"7b6eb987-454b-4c50-bb39-ede2e1006808\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.009631 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br9rq\" (UniqueName: \"kubernetes.io/projected/7b6eb987-454b-4c50-bb39-ede2e1006808-kube-api-access-br9rq\") pod \"7b6eb987-454b-4c50-bb39-ede2e1006808\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.010752 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-utilities" (OuterVolumeSpecName: "utilities") pod "7b6eb987-454b-4c50-bb39-ede2e1006808" (UID: "7b6eb987-454b-4c50-bb39-ede2e1006808"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.020546 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b6eb987-454b-4c50-bb39-ede2e1006808-kube-api-access-br9rq" (OuterVolumeSpecName: "kube-api-access-br9rq") pod "7b6eb987-454b-4c50-bb39-ede2e1006808" (UID: "7b6eb987-454b-4c50-bb39-ede2e1006808"). InnerVolumeSpecName "kube-api-access-br9rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.110412 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrrwv\" (UniqueName: \"kubernetes.io/projected/c963ddba-b620-46b4-83af-dfbaa96bbd03-kube-api-access-vrrwv\") pod \"c963ddba-b620-46b4-83af-dfbaa96bbd03\" (UID: \"c963ddba-b620-46b4-83af-dfbaa96bbd03\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.110524 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c963ddba-b620-46b4-83af-dfbaa96bbd03-catalog-content\") pod \"c963ddba-b620-46b4-83af-dfbaa96bbd03\" (UID: \"c963ddba-b620-46b4-83af-dfbaa96bbd03\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.110645 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b6eb987-454b-4c50-bb39-ede2e1006808" (UID: "7b6eb987-454b-4c50-bb39-ede2e1006808"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.110678 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c963ddba-b620-46b4-83af-dfbaa96bbd03-utilities\") pod \"c963ddba-b620-46b4-83af-dfbaa96bbd03\" (UID: \"c963ddba-b620-46b4-83af-dfbaa96bbd03\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.110888 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-catalog-content\") pod \"7b6eb987-454b-4c50-bb39-ede2e1006808\" (UID: \"7b6eb987-454b-4c50-bb39-ede2e1006808\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.111703 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c963ddba-b620-46b4-83af-dfbaa96bbd03-utilities" (OuterVolumeSpecName: "utilities") pod "c963ddba-b620-46b4-83af-dfbaa96bbd03" (UID: "c963ddba-b620-46b4-83af-dfbaa96bbd03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.111909 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.111993 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br9rq\" (UniqueName: \"kubernetes.io/projected/7b6eb987-454b-4c50-bb39-ede2e1006808-kube-api-access-br9rq\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: W0126 09:10:54.112004 4872 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/7b6eb987-454b-4c50-bb39-ede2e1006808/volumes/kubernetes.io~empty-dir/catalog-content Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.112020 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b6eb987-454b-4c50-bb39-ede2e1006808" (UID: "7b6eb987-454b-4c50-bb39-ede2e1006808"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.117366 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c963ddba-b620-46b4-83af-dfbaa96bbd03-kube-api-access-vrrwv" (OuterVolumeSpecName: "kube-api-access-vrrwv") pod "c963ddba-b620-46b4-83af-dfbaa96bbd03" (UID: "c963ddba-b620-46b4-83af-dfbaa96bbd03"). InnerVolumeSpecName "kube-api-access-vrrwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.213841 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrrwv\" (UniqueName: \"kubernetes.io/projected/c963ddba-b620-46b4-83af-dfbaa96bbd03-kube-api-access-vrrwv\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.214024 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c963ddba-b620-46b4-83af-dfbaa96bbd03-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.214055 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b6eb987-454b-4c50-bb39-ede2e1006808-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.227624 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c963ddba-b620-46b4-83af-dfbaa96bbd03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c963ddba-b620-46b4-83af-dfbaa96bbd03" (UID: "c963ddba-b620-46b4-83af-dfbaa96bbd03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.267832 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.315258 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zdlw\" (UniqueName: \"kubernetes.io/projected/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-kube-api-access-7zdlw\") pod \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\" (UID: \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.315311 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-marketplace-trusted-ca\") pod \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\" (UID: \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.315355 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-marketplace-operator-metrics\") pod \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\" (UID: \"07b7e8f0-a421-4fbc-9c0d-aa773210edcd\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.315623 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c963ddba-b620-46b4-83af-dfbaa96bbd03-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.316832 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "07b7e8f0-a421-4fbc-9c0d-aa773210edcd" (UID: "07b7e8f0-a421-4fbc-9c0d-aa773210edcd"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.318880 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-kube-api-access-7zdlw" (OuterVolumeSpecName: "kube-api-access-7zdlw") pod "07b7e8f0-a421-4fbc-9c0d-aa773210edcd" (UID: "07b7e8f0-a421-4fbc-9c0d-aa773210edcd"). InnerVolumeSpecName "kube-api-access-7zdlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.323002 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "07b7e8f0-a421-4fbc-9c0d-aa773210edcd" (UID: "07b7e8f0-a421-4fbc-9c0d-aa773210edcd"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.340579 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.416093 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48sxz\" (UniqueName: \"kubernetes.io/projected/3132ab11-2629-4089-b4e5-da75a4d83fd9-kube-api-access-48sxz\") pod \"3132ab11-2629-4089-b4e5-da75a4d83fd9\" (UID: \"3132ab11-2629-4089-b4e5-da75a4d83fd9\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.416141 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3132ab11-2629-4089-b4e5-da75a4d83fd9-catalog-content\") pod \"3132ab11-2629-4089-b4e5-da75a4d83fd9\" (UID: \"3132ab11-2629-4089-b4e5-da75a4d83fd9\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.416212 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3132ab11-2629-4089-b4e5-da75a4d83fd9-utilities\") pod \"3132ab11-2629-4089-b4e5-da75a4d83fd9\" (UID: \"3132ab11-2629-4089-b4e5-da75a4d83fd9\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.416650 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zdlw\" (UniqueName: \"kubernetes.io/projected/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-kube-api-access-7zdlw\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.416672 4872 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.416683 4872 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/07b7e8f0-a421-4fbc-9c0d-aa773210edcd-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.417371 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3132ab11-2629-4089-b4e5-da75a4d83fd9-utilities" (OuterVolumeSpecName: "utilities") pod "3132ab11-2629-4089-b4e5-da75a4d83fd9" (UID: "3132ab11-2629-4089-b4e5-da75a4d83fd9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.418534 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.420328 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3132ab11-2629-4089-b4e5-da75a4d83fd9-kube-api-access-48sxz" (OuterVolumeSpecName: "kube-api-access-48sxz") pod "3132ab11-2629-4089-b4e5-da75a4d83fd9" (UID: "3132ab11-2629-4089-b4e5-da75a4d83fd9"). InnerVolumeSpecName "kube-api-access-48sxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.455444 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3132ab11-2629-4089-b4e5-da75a4d83fd9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3132ab11-2629-4089-b4e5-da75a4d83fd9" (UID: "3132ab11-2629-4089-b4e5-da75a4d83fd9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.518824 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdf9w\" (UniqueName: \"kubernetes.io/projected/eedb2ebd-5684-464d-bc70-fba04fb6edbd-kube-api-access-qdf9w\") pod \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\" (UID: \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.518921 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eedb2ebd-5684-464d-bc70-fba04fb6edbd-utilities\") pod \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\" (UID: \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.518998 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eedb2ebd-5684-464d-bc70-fba04fb6edbd-catalog-content\") pod \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\" (UID: \"eedb2ebd-5684-464d-bc70-fba04fb6edbd\") " Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.519961 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eedb2ebd-5684-464d-bc70-fba04fb6edbd-utilities" (OuterVolumeSpecName: "utilities") pod "eedb2ebd-5684-464d-bc70-fba04fb6edbd" (UID: "eedb2ebd-5684-464d-bc70-fba04fb6edbd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.520064 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3132ab11-2629-4089-b4e5-da75a4d83fd9-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.520090 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48sxz\" (UniqueName: \"kubernetes.io/projected/3132ab11-2629-4089-b4e5-da75a4d83fd9-kube-api-access-48sxz\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.520105 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3132ab11-2629-4089-b4e5-da75a4d83fd9-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.522105 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eedb2ebd-5684-464d-bc70-fba04fb6edbd-kube-api-access-qdf9w" (OuterVolumeSpecName: "kube-api-access-qdf9w") pod "eedb2ebd-5684-464d-bc70-fba04fb6edbd" (UID: "eedb2ebd-5684-464d-bc70-fba04fb6edbd"). InnerVolumeSpecName "kube-api-access-qdf9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.618309 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5xspp" event={"ID":"7b6eb987-454b-4c50-bb39-ede2e1006808","Type":"ContainerDied","Data":"591d31cd59bc4cad588c4b8fa92c06297adbfb9f7f2af9e1689bd75165378942"} Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.618384 4872 scope.go:117] "RemoveContainer" containerID="f2d39ac93259445624371a26a04d3f7afed05a1b2b4838aceac1951edf35426a" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.618546 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5xspp" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.620469 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" event={"ID":"579725e4-42fb-49f8-b445-4eed08a373b6","Type":"ContainerStarted","Data":"5f6b13d8af5012d37680bf260decc2c5ad2625afa48fcecb05be35f4cbd69cc3"} Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.620523 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" event={"ID":"579725e4-42fb-49f8-b445-4eed08a373b6","Type":"ContainerStarted","Data":"fdc1486d28b0568fa6335022262f526d183b0c91444a169229f3e49742a6c46d"} Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.620838 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdf9w\" (UniqueName: \"kubernetes.io/projected/eedb2ebd-5684-464d-bc70-fba04fb6edbd-kube-api-access-qdf9w\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.620878 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.621317 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eedb2ebd-5684-464d-bc70-fba04fb6edbd-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.626014 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-84qrq" event={"ID":"3132ab11-2629-4089-b4e5-da75a4d83fd9","Type":"ContainerDied","Data":"3159a5365b4a224d049d3d2c32bc0ffaaa1fb28b6f22562b6d76a5b9610b1be5"} Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.626113 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-84qrq" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.628343 4872 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-9ckvb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" start-of-body= Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.628392 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" podUID="579725e4-42fb-49f8-b445-4eed08a373b6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.628782 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" event={"ID":"07b7e8f0-a421-4fbc-9c0d-aa773210edcd","Type":"ContainerDied","Data":"b1d1246dd553dbdaccfa0ceed7f1dcbacd99d48b9c2dd373fc9243873f273591"} Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.628906 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bsfdw" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.633863 4872 generic.go:334] "Generic (PLEG): container finished" podID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" containerID="eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c" exitCode=0 Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.633938 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cwfr" event={"ID":"eedb2ebd-5684-464d-bc70-fba04fb6edbd","Type":"ContainerDied","Data":"eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c"} Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.633975 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cwfr" event={"ID":"eedb2ebd-5684-464d-bc70-fba04fb6edbd","Type":"ContainerDied","Data":"4a0543754d2361565b10a92631d8b96118106db0bea4268ab7601df3d8990d4e"} Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.636169 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cwfr" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.641512 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p76kx" event={"ID":"c963ddba-b620-46b4-83af-dfbaa96bbd03","Type":"ContainerDied","Data":"b2a83f96f6f7d8a3a38b1aac0e94d08e82a8e2bc7bec8afaba3890b50f369993"} Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.641961 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p76kx" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.647233 4872 scope.go:117] "RemoveContainer" containerID="759ad08f8a02bf6c6af28abcbb2033a94c818e363d07b748bf476343b8ee23c6" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.649858 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" podStartSLOduration=1.649834638 podStartE2EDuration="1.649834638s" podCreationTimestamp="2026-01-26 09:10:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:10:54.648229823 +0000 UTC m=+187.957069624" watchObservedRunningTime="2026-01-26 09:10:54.649834638 +0000 UTC m=+187.958674439" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.650274 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eedb2ebd-5684-464d-bc70-fba04fb6edbd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eedb2ebd-5684-464d-bc70-fba04fb6edbd" (UID: "eedb2ebd-5684-464d-bc70-fba04fb6edbd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.686733 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bsfdw"] Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.690835 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bsfdw"] Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.692573 4872 scope.go:117] "RemoveContainer" containerID="433c56cbf0956956461b5d45447e7d80f6370c91b24a8cc388b45707a36112f3" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.706952 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-84qrq"] Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.724422 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eedb2ebd-5684-464d-bc70-fba04fb6edbd-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.733943 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-84qrq"] Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.757061 4872 scope.go:117] "RemoveContainer" containerID="f5b4fa1964841217ba62256094305892512dda59254f7b3046755749a7addadf" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.760222 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5xspp"] Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.763703 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5xspp"] Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.766474 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p76kx"] Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.768944 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p76kx"] Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.782393 4872 scope.go:117] "RemoveContainer" containerID="e5670ba6e88f896aee3a27ccec89990a2ab2556dce15db873a5622043e12c7fc" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.805100 4872 scope.go:117] "RemoveContainer" containerID="19e7a143bdb0f3e88e3be99e0b7f23e6960a8df3b681cd2e00b741f197ef2695" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.826014 4872 scope.go:117] "RemoveContainer" containerID="0f074f626e397d4d0e1b5b6e4c1aecb86c1a34078c0ce317ae30c6ebcfba7e34" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.845443 4872 scope.go:117] "RemoveContainer" containerID="eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.865541 4872 scope.go:117] "RemoveContainer" containerID="e31b0f5619db5b83f281302bcc70d512b01cfc4d5d9d13dd09f9df615c55707c" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.884353 4872 scope.go:117] "RemoveContainer" containerID="f10dba94e9487ea0c17a2c4068795ef39906b3cbbea831beb19e875132123eae" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.906388 4872 scope.go:117] "RemoveContainer" containerID="eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c" Jan 26 09:10:54 crc kubenswrapper[4872]: E0126 09:10:54.906949 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c\": container with ID starting with eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c not found: ID does not exist" containerID="eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.906985 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c"} err="failed to get container status \"eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c\": rpc error: code = NotFound desc = could not find container \"eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c\": container with ID starting with eb75cd0bf3d7f8202ac0dcd891ca06c7d272fd1be2e6edcded88d0989ff4ef2c not found: ID does not exist" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.907009 4872 scope.go:117] "RemoveContainer" containerID="e31b0f5619db5b83f281302bcc70d512b01cfc4d5d9d13dd09f9df615c55707c" Jan 26 09:10:54 crc kubenswrapper[4872]: E0126 09:10:54.907278 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e31b0f5619db5b83f281302bcc70d512b01cfc4d5d9d13dd09f9df615c55707c\": container with ID starting with e31b0f5619db5b83f281302bcc70d512b01cfc4d5d9d13dd09f9df615c55707c not found: ID does not exist" containerID="e31b0f5619db5b83f281302bcc70d512b01cfc4d5d9d13dd09f9df615c55707c" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.907293 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e31b0f5619db5b83f281302bcc70d512b01cfc4d5d9d13dd09f9df615c55707c"} err="failed to get container status \"e31b0f5619db5b83f281302bcc70d512b01cfc4d5d9d13dd09f9df615c55707c\": rpc error: code = NotFound desc = could not find container \"e31b0f5619db5b83f281302bcc70d512b01cfc4d5d9d13dd09f9df615c55707c\": container with ID starting with e31b0f5619db5b83f281302bcc70d512b01cfc4d5d9d13dd09f9df615c55707c not found: ID does not exist" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.907306 4872 scope.go:117] "RemoveContainer" containerID="f10dba94e9487ea0c17a2c4068795ef39906b3cbbea831beb19e875132123eae" Jan 26 09:10:54 crc kubenswrapper[4872]: E0126 09:10:54.907544 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f10dba94e9487ea0c17a2c4068795ef39906b3cbbea831beb19e875132123eae\": container with ID starting with f10dba94e9487ea0c17a2c4068795ef39906b3cbbea831beb19e875132123eae not found: ID does not exist" containerID="f10dba94e9487ea0c17a2c4068795ef39906b3cbbea831beb19e875132123eae" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.907559 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f10dba94e9487ea0c17a2c4068795ef39906b3cbbea831beb19e875132123eae"} err="failed to get container status \"f10dba94e9487ea0c17a2c4068795ef39906b3cbbea831beb19e875132123eae\": rpc error: code = NotFound desc = could not find container \"f10dba94e9487ea0c17a2c4068795ef39906b3cbbea831beb19e875132123eae\": container with ID starting with f10dba94e9487ea0c17a2c4068795ef39906b3cbbea831beb19e875132123eae not found: ID does not exist" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.907573 4872 scope.go:117] "RemoveContainer" containerID="3fc9adbf84cdcfe30525d46ca6e8e569b105295db734c5e3ad6d0709dded915b" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.924576 4872 scope.go:117] "RemoveContainer" containerID="d3df9b0d0f0530aca8ab3604ceae98306edc4cc2cf06f933d1859963c7525c06" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.943003 4872 scope.go:117] "RemoveContainer" containerID="1f9ac8ebee536943cb2adc8ff1454826058ea508b69dd0568adcd740c1a43cf6" Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.968264 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2cwfr"] Jan 26 09:10:54 crc kubenswrapper[4872]: I0126 09:10:54.974664 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2cwfr"] Jan 26 09:10:55 crc kubenswrapper[4872]: I0126 09:10:55.191989 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07b7e8f0-a421-4fbc-9c0d-aa773210edcd" path="/var/lib/kubelet/pods/07b7e8f0-a421-4fbc-9c0d-aa773210edcd/volumes" Jan 26 09:10:55 crc kubenswrapper[4872]: I0126 09:10:55.193762 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3132ab11-2629-4089-b4e5-da75a4d83fd9" path="/var/lib/kubelet/pods/3132ab11-2629-4089-b4e5-da75a4d83fd9/volumes" Jan 26 09:10:55 crc kubenswrapper[4872]: I0126 09:10:55.194528 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b6eb987-454b-4c50-bb39-ede2e1006808" path="/var/lib/kubelet/pods/7b6eb987-454b-4c50-bb39-ede2e1006808/volumes" Jan 26 09:10:55 crc kubenswrapper[4872]: I0126 09:10:55.195291 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c963ddba-b620-46b4-83af-dfbaa96bbd03" path="/var/lib/kubelet/pods/c963ddba-b620-46b4-83af-dfbaa96bbd03/volumes" Jan 26 09:10:55 crc kubenswrapper[4872]: I0126 09:10:55.196771 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" path="/var/lib/kubelet/pods/eedb2ebd-5684-464d-bc70-fba04fb6edbd/volumes" Jan 26 09:10:55 crc kubenswrapper[4872]: I0126 09:10:55.656926 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-9ckvb" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.299745 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rjj7m"] Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.300053 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" containerName="extract-utilities" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.300073 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" containerName="extract-utilities" Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.300091 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c963ddba-b620-46b4-83af-dfbaa96bbd03" containerName="extract-content" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.300102 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c963ddba-b620-46b4-83af-dfbaa96bbd03" containerName="extract-content" Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.300112 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" containerName="registry-server" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.300119 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" containerName="registry-server" Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.300440 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3132ab11-2629-4089-b4e5-da75a4d83fd9" containerName="extract-content" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.300455 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3132ab11-2629-4089-b4e5-da75a4d83fd9" containerName="extract-content" Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.300468 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3132ab11-2629-4089-b4e5-da75a4d83fd9" containerName="extract-utilities" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.300475 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3132ab11-2629-4089-b4e5-da75a4d83fd9" containerName="extract-utilities" Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.300485 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b6eb987-454b-4c50-bb39-ede2e1006808" containerName="registry-server" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.300492 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b6eb987-454b-4c50-bb39-ede2e1006808" containerName="registry-server" Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.300511 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c963ddba-b620-46b4-83af-dfbaa96bbd03" containerName="registry-server" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.300518 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c963ddba-b620-46b4-83af-dfbaa96bbd03" containerName="registry-server" Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.300528 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b6eb987-454b-4c50-bb39-ede2e1006808" containerName="extract-utilities" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.300535 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b6eb987-454b-4c50-bb39-ede2e1006808" containerName="extract-utilities" Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.300546 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c963ddba-b620-46b4-83af-dfbaa96bbd03" containerName="extract-utilities" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.300555 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c963ddba-b620-46b4-83af-dfbaa96bbd03" containerName="extract-utilities" Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.300564 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3132ab11-2629-4089-b4e5-da75a4d83fd9" containerName="registry-server" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.300572 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3132ab11-2629-4089-b4e5-da75a4d83fd9" containerName="registry-server" Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.301095 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b6eb987-454b-4c50-bb39-ede2e1006808" containerName="extract-content" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.301112 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b6eb987-454b-4c50-bb39-ede2e1006808" containerName="extract-content" Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.301263 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" containerName="extract-content" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.301270 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" containerName="extract-content" Jan 26 09:10:56 crc kubenswrapper[4872]: E0126 09:10:56.301278 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07b7e8f0-a421-4fbc-9c0d-aa773210edcd" containerName="marketplace-operator" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.301284 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="07b7e8f0-a421-4fbc-9c0d-aa773210edcd" containerName="marketplace-operator" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.301394 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="eedb2ebd-5684-464d-bc70-fba04fb6edbd" containerName="registry-server" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.301406 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b6eb987-454b-4c50-bb39-ede2e1006808" containerName="registry-server" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.301415 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="07b7e8f0-a421-4fbc-9c0d-aa773210edcd" containerName="marketplace-operator" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.301428 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="c963ddba-b620-46b4-83af-dfbaa96bbd03" containerName="registry-server" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.301439 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3132ab11-2629-4089-b4e5-da75a4d83fd9" containerName="registry-server" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.302350 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.308918 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.313074 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rjj7m"] Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.347775 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faac185b-bdff-4f6d-801a-1458123a20b0-utilities\") pod \"community-operators-rjj7m\" (UID: \"faac185b-bdff-4f6d-801a-1458123a20b0\") " pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.348071 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnnsn\" (UniqueName: \"kubernetes.io/projected/faac185b-bdff-4f6d-801a-1458123a20b0-kube-api-access-qnnsn\") pod \"community-operators-rjj7m\" (UID: \"faac185b-bdff-4f6d-801a-1458123a20b0\") " pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.348211 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faac185b-bdff-4f6d-801a-1458123a20b0-catalog-content\") pod \"community-operators-rjj7m\" (UID: \"faac185b-bdff-4f6d-801a-1458123a20b0\") " pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.449323 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnnsn\" (UniqueName: \"kubernetes.io/projected/faac185b-bdff-4f6d-801a-1458123a20b0-kube-api-access-qnnsn\") pod \"community-operators-rjj7m\" (UID: \"faac185b-bdff-4f6d-801a-1458123a20b0\") " pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.449404 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faac185b-bdff-4f6d-801a-1458123a20b0-catalog-content\") pod \"community-operators-rjj7m\" (UID: \"faac185b-bdff-4f6d-801a-1458123a20b0\") " pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.449443 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faac185b-bdff-4f6d-801a-1458123a20b0-utilities\") pod \"community-operators-rjj7m\" (UID: \"faac185b-bdff-4f6d-801a-1458123a20b0\") " pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.450108 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/faac185b-bdff-4f6d-801a-1458123a20b0-catalog-content\") pod \"community-operators-rjj7m\" (UID: \"faac185b-bdff-4f6d-801a-1458123a20b0\") " pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.450415 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/faac185b-bdff-4f6d-801a-1458123a20b0-utilities\") pod \"community-operators-rjj7m\" (UID: \"faac185b-bdff-4f6d-801a-1458123a20b0\") " pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.470938 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnnsn\" (UniqueName: \"kubernetes.io/projected/faac185b-bdff-4f6d-801a-1458123a20b0-kube-api-access-qnnsn\") pod \"community-operators-rjj7m\" (UID: \"faac185b-bdff-4f6d-801a-1458123a20b0\") " pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.626622 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.903313 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d5qrt"] Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.904962 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.909270 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.914298 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d5qrt"] Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.957014 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fccd3bc2-18ff-40d4-b862-7257f6cebc5b-catalog-content\") pod \"redhat-operators-d5qrt\" (UID: \"fccd3bc2-18ff-40d4-b862-7257f6cebc5b\") " pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.957130 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-th8xv\" (UniqueName: \"kubernetes.io/projected/fccd3bc2-18ff-40d4-b862-7257f6cebc5b-kube-api-access-th8xv\") pod \"redhat-operators-d5qrt\" (UID: \"fccd3bc2-18ff-40d4-b862-7257f6cebc5b\") " pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:10:56 crc kubenswrapper[4872]: I0126 09:10:56.957187 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fccd3bc2-18ff-40d4-b862-7257f6cebc5b-utilities\") pod \"redhat-operators-d5qrt\" (UID: \"fccd3bc2-18ff-40d4-b862-7257f6cebc5b\") " pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.058631 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fccd3bc2-18ff-40d4-b862-7257f6cebc5b-utilities\") pod \"redhat-operators-d5qrt\" (UID: \"fccd3bc2-18ff-40d4-b862-7257f6cebc5b\") " pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.058742 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fccd3bc2-18ff-40d4-b862-7257f6cebc5b-catalog-content\") pod \"redhat-operators-d5qrt\" (UID: \"fccd3bc2-18ff-40d4-b862-7257f6cebc5b\") " pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.058785 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-th8xv\" (UniqueName: \"kubernetes.io/projected/fccd3bc2-18ff-40d4-b862-7257f6cebc5b-kube-api-access-th8xv\") pod \"redhat-operators-d5qrt\" (UID: \"fccd3bc2-18ff-40d4-b862-7257f6cebc5b\") " pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.059236 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fccd3bc2-18ff-40d4-b862-7257f6cebc5b-utilities\") pod \"redhat-operators-d5qrt\" (UID: \"fccd3bc2-18ff-40d4-b862-7257f6cebc5b\") " pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.059272 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fccd3bc2-18ff-40d4-b862-7257f6cebc5b-catalog-content\") pod \"redhat-operators-d5qrt\" (UID: \"fccd3bc2-18ff-40d4-b862-7257f6cebc5b\") " pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.073487 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rjj7m"] Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.077715 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-th8xv\" (UniqueName: \"kubernetes.io/projected/fccd3bc2-18ff-40d4-b862-7257f6cebc5b-kube-api-access-th8xv\") pod \"redhat-operators-d5qrt\" (UID: \"fccd3bc2-18ff-40d4-b862-7257f6cebc5b\") " pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:10:57 crc kubenswrapper[4872]: W0126 09:10:57.084087 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfaac185b_bdff_4f6d_801a_1458123a20b0.slice/crio-a59dd8e011a81e19fd34a69eceaa5e2fa5fda5ccfb36aa053f8a0d5e7c9aa160 WatchSource:0}: Error finding container a59dd8e011a81e19fd34a69eceaa5e2fa5fda5ccfb36aa053f8a0d5e7c9aa160: Status 404 returned error can't find the container with id a59dd8e011a81e19fd34a69eceaa5e2fa5fda5ccfb36aa053f8a0d5e7c9aa160 Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.229233 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.466777 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d5qrt"] Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.662996 4872 generic.go:334] "Generic (PLEG): container finished" podID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" containerID="e3c3bdf40099d6c5416344c0e351c8b7a80179eb0efd79117b82ac12901494ea" exitCode=0 Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.663240 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d5qrt" event={"ID":"fccd3bc2-18ff-40d4-b862-7257f6cebc5b","Type":"ContainerDied","Data":"e3c3bdf40099d6c5416344c0e351c8b7a80179eb0efd79117b82ac12901494ea"} Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.663560 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d5qrt" event={"ID":"fccd3bc2-18ff-40d4-b862-7257f6cebc5b","Type":"ContainerStarted","Data":"677c1bad5be0babf5adbc4907772592987d7fb81b24fc4459b08599b90aa97e1"} Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.672172 4872 generic.go:334] "Generic (PLEG): container finished" podID="faac185b-bdff-4f6d-801a-1458123a20b0" containerID="f988152f9fa6d513bd730d61d96b9fb7fd3da03261e387ea8f6e156a45eb401e" exitCode=0 Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.672215 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjj7m" event={"ID":"faac185b-bdff-4f6d-801a-1458123a20b0","Type":"ContainerDied","Data":"f988152f9fa6d513bd730d61d96b9fb7fd3da03261e387ea8f6e156a45eb401e"} Jan 26 09:10:57 crc kubenswrapper[4872]: I0126 09:10:57.672247 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjj7m" event={"ID":"faac185b-bdff-4f6d-801a-1458123a20b0","Type":"ContainerStarted","Data":"a59dd8e011a81e19fd34a69eceaa5e2fa5fda5ccfb36aa053f8a0d5e7c9aa160"} Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.696141 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5mk2s"] Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.697706 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.700992 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.711133 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5mk2s"] Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.720074 4872 generic.go:334] "Generic (PLEG): container finished" podID="faac185b-bdff-4f6d-801a-1458123a20b0" containerID="5598908f011feb7e6295b07e8b3f561cf74a203e60896682024d96f1f36c6872" exitCode=0 Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.720152 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjj7m" event={"ID":"faac185b-bdff-4f6d-801a-1458123a20b0","Type":"ContainerDied","Data":"5598908f011feb7e6295b07e8b3f561cf74a203e60896682024d96f1f36c6872"} Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.795160 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f25adda-7523-4fba-9f9e-32583f0d9ad6-utilities\") pod \"certified-operators-5mk2s\" (UID: \"5f25adda-7523-4fba-9f9e-32583f0d9ad6\") " pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.795223 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f25adda-7523-4fba-9f9e-32583f0d9ad6-catalog-content\") pod \"certified-operators-5mk2s\" (UID: \"5f25adda-7523-4fba-9f9e-32583f0d9ad6\") " pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.795284 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw7vz\" (UniqueName: \"kubernetes.io/projected/5f25adda-7523-4fba-9f9e-32583f0d9ad6-kube-api-access-tw7vz\") pod \"certified-operators-5mk2s\" (UID: \"5f25adda-7523-4fba-9f9e-32583f0d9ad6\") " pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.897071 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw7vz\" (UniqueName: \"kubernetes.io/projected/5f25adda-7523-4fba-9f9e-32583f0d9ad6-kube-api-access-tw7vz\") pod \"certified-operators-5mk2s\" (UID: \"5f25adda-7523-4fba-9f9e-32583f0d9ad6\") " pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.897182 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f25adda-7523-4fba-9f9e-32583f0d9ad6-utilities\") pod \"certified-operators-5mk2s\" (UID: \"5f25adda-7523-4fba-9f9e-32583f0d9ad6\") " pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.897212 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f25adda-7523-4fba-9f9e-32583f0d9ad6-catalog-content\") pod \"certified-operators-5mk2s\" (UID: \"5f25adda-7523-4fba-9f9e-32583f0d9ad6\") " pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.897657 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5f25adda-7523-4fba-9f9e-32583f0d9ad6-catalog-content\") pod \"certified-operators-5mk2s\" (UID: \"5f25adda-7523-4fba-9f9e-32583f0d9ad6\") " pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.897716 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5f25adda-7523-4fba-9f9e-32583f0d9ad6-utilities\") pod \"certified-operators-5mk2s\" (UID: \"5f25adda-7523-4fba-9f9e-32583f0d9ad6\") " pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:10:58 crc kubenswrapper[4872]: I0126 09:10:58.919794 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw7vz\" (UniqueName: \"kubernetes.io/projected/5f25adda-7523-4fba-9f9e-32583f0d9ad6-kube-api-access-tw7vz\") pod \"certified-operators-5mk2s\" (UID: \"5f25adda-7523-4fba-9f9e-32583f0d9ad6\") " pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.024900 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.307506 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rnmnc"] Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.309349 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.313060 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.316099 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rnmnc"] Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.409417 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgpx7\" (UniqueName: \"kubernetes.io/projected/41ca443e-2b21-4369-851c-30b312340df2-kube-api-access-pgpx7\") pod \"redhat-marketplace-rnmnc\" (UID: \"41ca443e-2b21-4369-851c-30b312340df2\") " pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.409496 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41ca443e-2b21-4369-851c-30b312340df2-catalog-content\") pod \"redhat-marketplace-rnmnc\" (UID: \"41ca443e-2b21-4369-851c-30b312340df2\") " pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.409614 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41ca443e-2b21-4369-851c-30b312340df2-utilities\") pod \"redhat-marketplace-rnmnc\" (UID: \"41ca443e-2b21-4369-851c-30b312340df2\") " pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.456718 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5mk2s"] Jan 26 09:10:59 crc kubenswrapper[4872]: W0126 09:10:59.473223 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f25adda_7523_4fba_9f9e_32583f0d9ad6.slice/crio-538d03d2ab3e211500cbd0d4dec01dae63658e9c861743ab669d846b1d63b8cc WatchSource:0}: Error finding container 538d03d2ab3e211500cbd0d4dec01dae63658e9c861743ab669d846b1d63b8cc: Status 404 returned error can't find the container with id 538d03d2ab3e211500cbd0d4dec01dae63658e9c861743ab669d846b1d63b8cc Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.513428 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgpx7\" (UniqueName: \"kubernetes.io/projected/41ca443e-2b21-4369-851c-30b312340df2-kube-api-access-pgpx7\") pod \"redhat-marketplace-rnmnc\" (UID: \"41ca443e-2b21-4369-851c-30b312340df2\") " pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.513481 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41ca443e-2b21-4369-851c-30b312340df2-catalog-content\") pod \"redhat-marketplace-rnmnc\" (UID: \"41ca443e-2b21-4369-851c-30b312340df2\") " pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.513501 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41ca443e-2b21-4369-851c-30b312340df2-utilities\") pod \"redhat-marketplace-rnmnc\" (UID: \"41ca443e-2b21-4369-851c-30b312340df2\") " pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.514082 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/41ca443e-2b21-4369-851c-30b312340df2-utilities\") pod \"redhat-marketplace-rnmnc\" (UID: \"41ca443e-2b21-4369-851c-30b312340df2\") " pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.514312 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/41ca443e-2b21-4369-851c-30b312340df2-catalog-content\") pod \"redhat-marketplace-rnmnc\" (UID: \"41ca443e-2b21-4369-851c-30b312340df2\") " pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.534996 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgpx7\" (UniqueName: \"kubernetes.io/projected/41ca443e-2b21-4369-851c-30b312340df2-kube-api-access-pgpx7\") pod \"redhat-marketplace-rnmnc\" (UID: \"41ca443e-2b21-4369-851c-30b312340df2\") " pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.644472 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.734981 4872 generic.go:334] "Generic (PLEG): container finished" podID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" containerID="edc675060d9264ea876b4c21ae21535d6cdad3a6521988cc15f31c49d4473468" exitCode=0 Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.735056 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d5qrt" event={"ID":"fccd3bc2-18ff-40d4-b862-7257f6cebc5b","Type":"ContainerDied","Data":"edc675060d9264ea876b4c21ae21535d6cdad3a6521988cc15f31c49d4473468"} Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.738267 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rjj7m" event={"ID":"faac185b-bdff-4f6d-801a-1458123a20b0","Type":"ContainerStarted","Data":"199cb854bd17510244e1200bcfc8e84926b1fbb414388528b6e8121f652f17e8"} Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.775395 4872 generic.go:334] "Generic (PLEG): container finished" podID="5f25adda-7523-4fba-9f9e-32583f0d9ad6" containerID="8fbc2d64ebeaea276f2112b7e24e91b611a9f89df1757c9008a9cf46b6ee9866" exitCode=0 Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.775513 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mk2s" event={"ID":"5f25adda-7523-4fba-9f9e-32583f0d9ad6","Type":"ContainerDied","Data":"8fbc2d64ebeaea276f2112b7e24e91b611a9f89df1757c9008a9cf46b6ee9866"} Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.775854 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mk2s" event={"ID":"5f25adda-7523-4fba-9f9e-32583f0d9ad6","Type":"ContainerStarted","Data":"538d03d2ab3e211500cbd0d4dec01dae63658e9c861743ab669d846b1d63b8cc"} Jan 26 09:10:59 crc kubenswrapper[4872]: I0126 09:10:59.792834 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rjj7m" podStartSLOduration=2.335758556 podStartE2EDuration="3.792778955s" podCreationTimestamp="2026-01-26 09:10:56 +0000 UTC" firstStartedPulling="2026-01-26 09:10:57.674454792 +0000 UTC m=+190.983294593" lastFinishedPulling="2026-01-26 09:10:59.131475191 +0000 UTC m=+192.440314992" observedRunningTime="2026-01-26 09:10:59.78968262 +0000 UTC m=+193.098522431" watchObservedRunningTime="2026-01-26 09:10:59.792778955 +0000 UTC m=+193.101618756" Jan 26 09:11:00 crc kubenswrapper[4872]: I0126 09:11:00.098659 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rnmnc"] Jan 26 09:11:00 crc kubenswrapper[4872]: I0126 09:11:00.784889 4872 generic.go:334] "Generic (PLEG): container finished" podID="41ca443e-2b21-4369-851c-30b312340df2" containerID="3f9e08e3f22c52e6262e45ae068fc0555877a22d63f5b1bc104eb574df2c3110" exitCode=0 Jan 26 09:11:00 crc kubenswrapper[4872]: I0126 09:11:00.784953 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnmnc" event={"ID":"41ca443e-2b21-4369-851c-30b312340df2","Type":"ContainerDied","Data":"3f9e08e3f22c52e6262e45ae068fc0555877a22d63f5b1bc104eb574df2c3110"} Jan 26 09:11:00 crc kubenswrapper[4872]: I0126 09:11:00.785425 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnmnc" event={"ID":"41ca443e-2b21-4369-851c-30b312340df2","Type":"ContainerStarted","Data":"e88be9c6ccd9d626dd8570fd740345c970a9cdd1f04723cbc47660cfea1024f4"} Jan 26 09:11:00 crc kubenswrapper[4872]: I0126 09:11:00.792530 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d5qrt" event={"ID":"fccd3bc2-18ff-40d4-b862-7257f6cebc5b","Type":"ContainerStarted","Data":"1ac106f2e88f25c85008bb701a42f239ad04c2d44ee4e24a1ee5a56d58940066"} Jan 26 09:11:00 crc kubenswrapper[4872]: I0126 09:11:00.795494 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mk2s" event={"ID":"5f25adda-7523-4fba-9f9e-32583f0d9ad6","Type":"ContainerStarted","Data":"fb7a1b990b5e07e907c29514185c15b5df33d6c0110f9aa84486a66f07d67b35"} Jan 26 09:11:00 crc kubenswrapper[4872]: I0126 09:11:00.856942 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d5qrt" podStartSLOduration=2.235890285 podStartE2EDuration="4.856910004s" podCreationTimestamp="2026-01-26 09:10:56 +0000 UTC" firstStartedPulling="2026-01-26 09:10:57.666113132 +0000 UTC m=+190.974952933" lastFinishedPulling="2026-01-26 09:11:00.287132851 +0000 UTC m=+193.595972652" observedRunningTime="2026-01-26 09:11:00.85422103 +0000 UTC m=+194.163060861" watchObservedRunningTime="2026-01-26 09:11:00.856910004 +0000 UTC m=+194.165749815" Jan 26 09:11:01 crc kubenswrapper[4872]: I0126 09:11:01.806431 4872 generic.go:334] "Generic (PLEG): container finished" podID="5f25adda-7523-4fba-9f9e-32583f0d9ad6" containerID="fb7a1b990b5e07e907c29514185c15b5df33d6c0110f9aa84486a66f07d67b35" exitCode=0 Jan 26 09:11:01 crc kubenswrapper[4872]: I0126 09:11:01.806587 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mk2s" event={"ID":"5f25adda-7523-4fba-9f9e-32583f0d9ad6","Type":"ContainerDied","Data":"fb7a1b990b5e07e907c29514185c15b5df33d6c0110f9aa84486a66f07d67b35"} Jan 26 09:11:01 crc kubenswrapper[4872]: I0126 09:11:01.810002 4872 generic.go:334] "Generic (PLEG): container finished" podID="41ca443e-2b21-4369-851c-30b312340df2" containerID="e62d1d5de4c6b95da98e671a9b3dc2d47da975d7136e36ded25cbd49c10e2646" exitCode=0 Jan 26 09:11:01 crc kubenswrapper[4872]: I0126 09:11:01.810928 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnmnc" event={"ID":"41ca443e-2b21-4369-851c-30b312340df2","Type":"ContainerDied","Data":"e62d1d5de4c6b95da98e671a9b3dc2d47da975d7136e36ded25cbd49c10e2646"} Jan 26 09:11:02 crc kubenswrapper[4872]: I0126 09:11:02.817934 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rnmnc" event={"ID":"41ca443e-2b21-4369-851c-30b312340df2","Type":"ContainerStarted","Data":"127afbf6891ff1da4195a3626d7ed1515c19119f2ff573e94ec1892335f60f94"} Jan 26 09:11:02 crc kubenswrapper[4872]: I0126 09:11:02.821188 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5mk2s" event={"ID":"5f25adda-7523-4fba-9f9e-32583f0d9ad6","Type":"ContainerStarted","Data":"f4909904a2fe81735dc13d4f3ebe50fe80583b045206f65a4e9cb4d99ea7ceef"} Jan 26 09:11:02 crc kubenswrapper[4872]: I0126 09:11:02.843120 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rnmnc" podStartSLOduration=1.994752979 podStartE2EDuration="3.843096997s" podCreationTimestamp="2026-01-26 09:10:59 +0000 UTC" firstStartedPulling="2026-01-26 09:11:00.787768489 +0000 UTC m=+194.096608290" lastFinishedPulling="2026-01-26 09:11:02.636112507 +0000 UTC m=+195.944952308" observedRunningTime="2026-01-26 09:11:02.839265652 +0000 UTC m=+196.148105453" watchObservedRunningTime="2026-01-26 09:11:02.843096997 +0000 UTC m=+196.151936798" Jan 26 09:11:02 crc kubenswrapper[4872]: I0126 09:11:02.867386 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5mk2s" podStartSLOduration=2.407909507 podStartE2EDuration="4.867360376s" podCreationTimestamp="2026-01-26 09:10:58 +0000 UTC" firstStartedPulling="2026-01-26 09:10:59.777944756 +0000 UTC m=+193.086784557" lastFinishedPulling="2026-01-26 09:11:02.237395625 +0000 UTC m=+195.546235426" observedRunningTime="2026-01-26 09:11:02.864643301 +0000 UTC m=+196.173483102" watchObservedRunningTime="2026-01-26 09:11:02.867360376 +0000 UTC m=+196.176200177" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.728975 4872 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.730345 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.731300 4872 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.731791 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd" gracePeriod=15 Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.731848 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4" gracePeriod=15 Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.731883 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091" gracePeriod=15 Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.731855 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2" gracePeriod=15 Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.731739 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150" gracePeriod=15 Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.734599 4872 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 09:11:04 crc kubenswrapper[4872]: E0126 09:11:04.734905 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.734930 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 26 09:11:04 crc kubenswrapper[4872]: E0126 09:11:04.734942 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.734951 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 26 09:11:04 crc kubenswrapper[4872]: E0126 09:11:04.734961 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.734968 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 09:11:04 crc kubenswrapper[4872]: E0126 09:11:04.734982 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.734988 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Jan 26 09:11:04 crc kubenswrapper[4872]: E0126 09:11:04.734998 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.735006 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 26 09:11:04 crc kubenswrapper[4872]: E0126 09:11:04.735016 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.735023 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 09:11:04 crc kubenswrapper[4872]: E0126 09:11:04.735034 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.735042 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.735261 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.735303 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.735315 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.735325 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.735371 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.735381 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.775451 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.809301 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.809394 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.809427 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.809446 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.809468 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.809487 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.809500 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.809525 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.910523 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.910701 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.910823 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.910897 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.910949 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.911000 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.910961 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.911117 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.911145 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.911233 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.911239 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.911294 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.911343 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.911406 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.911359 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:04 crc kubenswrapper[4872]: I0126 09:11:04.911547 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.064154 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:11:05 crc kubenswrapper[4872]: E0126 09:11:05.104982 4872 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.70:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188e3ce6b078abfa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-26 09:11:05.104026618 +0000 UTC m=+198.412866419,LastTimestamp:2026-01-26 09:11:05.104026618 +0000 UTC m=+198.412866419,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.864152 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.868643 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.870139 4872 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4" exitCode=0 Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.870214 4872 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091" exitCode=0 Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.870225 4872 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd" exitCode=0 Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.870237 4872 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2" exitCode=2 Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.870271 4872 scope.go:117] "RemoveContainer" containerID="ae37ff180f69e782dde9ca2d613dafbaa5756f12dc29aeaeabee560dabe42668" Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.875556 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"66126fdf89160b3f4d8fee22a92d4e517fc226b6a05cbae1286525e4e07373af"} Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.875630 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"fc960fc51839943dc1d5c849b10901e69d57b4d37b768ebab2344c0559f0cd5b"} Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.877537 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.878944 4872 generic.go:334] "Generic (PLEG): container finished" podID="717a9b10-82c6-4474-8715-0b34ff4e391e" containerID="01acc704dd63659f2b630abc8c9febe9217b337227f35452b306fbddd087f4ed" exitCode=0 Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.878983 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"717a9b10-82c6-4474-8715-0b34ff4e391e","Type":"ContainerDied","Data":"01acc704dd63659f2b630abc8c9febe9217b337227f35452b306fbddd087f4ed"} Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.879344 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:05 crc kubenswrapper[4872]: I0126 09:11:05.879540 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:06 crc kubenswrapper[4872]: I0126 09:11:06.627932 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:11:06 crc kubenswrapper[4872]: I0126 09:11:06.628456 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:11:06 crc kubenswrapper[4872]: I0126 09:11:06.676367 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:11:06 crc kubenswrapper[4872]: I0126 09:11:06.677119 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:06 crc kubenswrapper[4872]: I0126 09:11:06.677864 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:06 crc kubenswrapper[4872]: I0126 09:11:06.678369 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:06 crc kubenswrapper[4872]: I0126 09:11:06.891493 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 09:11:06 crc kubenswrapper[4872]: I0126 09:11:06.971929 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rjj7m" Jan 26 09:11:06 crc kubenswrapper[4872]: I0126 09:11:06.973141 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:06 crc kubenswrapper[4872]: I0126 09:11:06.973699 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:06 crc kubenswrapper[4872]: I0126 09:11:06.974158 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.188256 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.189599 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.189649 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.190108 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.190407 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.190603 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.190786 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.229922 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.230285 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.246652 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/717a9b10-82c6-4474-8715-0b34ff4e391e-kube-api-access\") pod \"717a9b10-82c6-4474-8715-0b34ff4e391e\" (UID: \"717a9b10-82c6-4474-8715-0b34ff4e391e\") " Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.247750 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/717a9b10-82c6-4474-8715-0b34ff4e391e-kubelet-dir\") pod \"717a9b10-82c6-4474-8715-0b34ff4e391e\" (UID: \"717a9b10-82c6-4474-8715-0b34ff4e391e\") " Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.247913 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/717a9b10-82c6-4474-8715-0b34ff4e391e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "717a9b10-82c6-4474-8715-0b34ff4e391e" (UID: "717a9b10-82c6-4474-8715-0b34ff4e391e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.247965 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/717a9b10-82c6-4474-8715-0b34ff4e391e-var-lock\") pod \"717a9b10-82c6-4474-8715-0b34ff4e391e\" (UID: \"717a9b10-82c6-4474-8715-0b34ff4e391e\") " Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.248673 4872 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/717a9b10-82c6-4474-8715-0b34ff4e391e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.248075 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/717a9b10-82c6-4474-8715-0b34ff4e391e-var-lock" (OuterVolumeSpecName: "var-lock") pod "717a9b10-82c6-4474-8715-0b34ff4e391e" (UID: "717a9b10-82c6-4474-8715-0b34ff4e391e"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.258576 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/717a9b10-82c6-4474-8715-0b34ff4e391e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "717a9b10-82c6-4474-8715-0b34ff4e391e" (UID: "717a9b10-82c6-4474-8715-0b34ff4e391e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.277106 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.278767 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.279791 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.281247 4872 status_manager.go:851] "Failed to get status for pod" podUID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" pod="openshift-marketplace/redhat-operators-d5qrt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d5qrt\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.281869 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.349509 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/717a9b10-82c6-4474-8715-0b34ff4e391e-kube-api-access\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.349545 4872 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/717a9b10-82c6-4474-8715-0b34ff4e391e-var-lock\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.628588 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.629696 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.630511 4872 status_manager.go:851] "Failed to get status for pod" podUID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" pod="openshift-marketplace/redhat-operators-d5qrt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d5qrt\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.630702 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.631095 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.631684 4872 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.632147 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.754554 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.754606 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.754725 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.755020 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.755057 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.755072 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.856353 4872 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.856400 4872 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.856414 4872 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.908144 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.908924 4872 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150" exitCode=0 Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.909002 4872 scope.go:117] "RemoveContainer" containerID="5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.909175 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.911654 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"717a9b10-82c6-4474-8715-0b34ff4e391e","Type":"ContainerDied","Data":"9ed1f368a28d82c5b6236341331dd0f01711b1eb8cdbe4dea14624463b933e96"} Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.911724 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ed1f368a28d82c5b6236341331dd0f01711b1eb8cdbe4dea14624463b933e96" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.911919 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.932376 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.932914 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.933393 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.934469 4872 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.934718 4872 status_manager.go:851] "Failed to get status for pod" podUID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" pod="openshift-marketplace/redhat-operators-d5qrt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d5qrt\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.935016 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.935329 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.935863 4872 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.936206 4872 status_manager.go:851] "Failed to get status for pod" podUID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" pod="openshift-marketplace/redhat-operators-d5qrt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d5qrt\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.936663 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.938445 4872 scope.go:117] "RemoveContainer" containerID="e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.961107 4872 scope.go:117] "RemoveContainer" containerID="fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.962714 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d5qrt" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.963489 4872 status_manager.go:851] "Failed to get status for pod" podUID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" pod="openshift-marketplace/redhat-operators-d5qrt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d5qrt\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.963684 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.964053 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.964359 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.964869 4872 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.979280 4872 scope.go:117] "RemoveContainer" containerID="198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2" Jan 26 09:11:07 crc kubenswrapper[4872]: I0126 09:11:07.994996 4872 scope.go:117] "RemoveContainer" containerID="95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.016618 4872 scope.go:117] "RemoveContainer" containerID="b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.041046 4872 scope.go:117] "RemoveContainer" containerID="5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4" Jan 26 09:11:08 crc kubenswrapper[4872]: E0126 09:11:08.041789 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\": container with ID starting with 5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4 not found: ID does not exist" containerID="5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.041885 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4"} err="failed to get container status \"5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\": rpc error: code = NotFound desc = could not find container \"5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4\": container with ID starting with 5cf6ec89ec97227091c20a9f923daa6cbf8f706aec7e09a78eb73646fd315dc4 not found: ID does not exist" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.041944 4872 scope.go:117] "RemoveContainer" containerID="e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091" Jan 26 09:11:08 crc kubenswrapper[4872]: E0126 09:11:08.047481 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\": container with ID starting with e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091 not found: ID does not exist" containerID="e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.047534 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091"} err="failed to get container status \"e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\": rpc error: code = NotFound desc = could not find container \"e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091\": container with ID starting with e9b971e4ebaab94e5cc6301f8c0c476735559cda5fbadee367bd24ede5259091 not found: ID does not exist" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.047576 4872 scope.go:117] "RemoveContainer" containerID="fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd" Jan 26 09:11:08 crc kubenswrapper[4872]: E0126 09:11:08.048159 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\": container with ID starting with fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd not found: ID does not exist" containerID="fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.048182 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd"} err="failed to get container status \"fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\": rpc error: code = NotFound desc = could not find container \"fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd\": container with ID starting with fe3217f76bbd28b1ebf2796b13d4aebb03d7e103bf28d55999e5fbd6e33939bd not found: ID does not exist" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.048194 4872 scope.go:117] "RemoveContainer" containerID="198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2" Jan 26 09:11:08 crc kubenswrapper[4872]: E0126 09:11:08.048439 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\": container with ID starting with 198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2 not found: ID does not exist" containerID="198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.048454 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2"} err="failed to get container status \"198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\": rpc error: code = NotFound desc = could not find container \"198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2\": container with ID starting with 198f8144880e5a2e747626685f9c541316a3ee811cdc6825f6790d805d4a6bb2 not found: ID does not exist" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.048470 4872 scope.go:117] "RemoveContainer" containerID="95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150" Jan 26 09:11:08 crc kubenswrapper[4872]: E0126 09:11:08.048789 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\": container with ID starting with 95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150 not found: ID does not exist" containerID="95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.048838 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150"} err="failed to get container status \"95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\": rpc error: code = NotFound desc = could not find container \"95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150\": container with ID starting with 95639005ad7ac6380c26cf4d7a519f86ce8f3c7e02930d4c6d7c71cc6d964150 not found: ID does not exist" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.048852 4872 scope.go:117] "RemoveContainer" containerID="b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785" Jan 26 09:11:08 crc kubenswrapper[4872]: E0126 09:11:08.049302 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\": container with ID starting with b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785 not found: ID does not exist" containerID="b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785" Jan 26 09:11:08 crc kubenswrapper[4872]: I0126 09:11:08.049356 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785"} err="failed to get container status \"b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\": rpc error: code = NotFound desc = could not find container \"b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785\": container with ID starting with b6e62015c182f9cc5dce0eabdd5c7c8bc1249cc4f90a11ec57b229ab38dbb785 not found: ID does not exist" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.025785 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.026294 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.074483 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.075303 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.075839 4872 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.076567 4872 status_manager.go:851] "Failed to get status for pod" podUID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" pod="openshift-marketplace/redhat-operators-d5qrt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d5qrt\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.077058 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.077671 4872 status_manager.go:851] "Failed to get status for pod" podUID="5f25adda-7523-4fba-9f9e-32583f0d9ad6" pod="openshift-marketplace/certified-operators-5mk2s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5mk2s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.078114 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.196313 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.645594 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.646079 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.698436 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.699169 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.700349 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.701079 4872 status_manager.go:851] "Failed to get status for pod" podUID="41ca443e-2b21-4369-851c-30b312340df2" pod="openshift-marketplace/redhat-marketplace-rnmnc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rnmnc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.701479 4872 status_manager.go:851] "Failed to get status for pod" podUID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" pod="openshift-marketplace/redhat-operators-d5qrt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d5qrt\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.701787 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.702176 4872 status_manager.go:851] "Failed to get status for pod" podUID="5f25adda-7523-4fba-9f9e-32583f0d9ad6" pod="openshift-marketplace/certified-operators-5mk2s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5mk2s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: E0126 09:11:09.961015 4872 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.70:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.188e3ce6b078abfa openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-01-26 09:11:05.104026618 +0000 UTC m=+198.412866419,LastTimestamp:2026-01-26 09:11:05.104026618 +0000 UTC m=+198.412866419,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.975462 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rnmnc" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.975958 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5mk2s" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.976193 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.976543 4872 status_manager.go:851] "Failed to get status for pod" podUID="41ca443e-2b21-4369-851c-30b312340df2" pod="openshift-marketplace/redhat-marketplace-rnmnc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rnmnc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.977006 4872 status_manager.go:851] "Failed to get status for pod" podUID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" pod="openshift-marketplace/redhat-operators-d5qrt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d5qrt\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.977257 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.977562 4872 status_manager.go:851] "Failed to get status for pod" podUID="5f25adda-7523-4fba-9f9e-32583f0d9ad6" pod="openshift-marketplace/certified-operators-5mk2s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5mk2s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.977855 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.978230 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.978509 4872 status_manager.go:851] "Failed to get status for pod" podUID="41ca443e-2b21-4369-851c-30b312340df2" pod="openshift-marketplace/redhat-marketplace-rnmnc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rnmnc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.978817 4872 status_manager.go:851] "Failed to get status for pod" podUID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" pod="openshift-marketplace/redhat-operators-d5qrt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d5qrt\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.979075 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.979312 4872 status_manager.go:851] "Failed to get status for pod" podUID="5f25adda-7523-4fba-9f9e-32583f0d9ad6" pod="openshift-marketplace/certified-operators-5mk2s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5mk2s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:09 crc kubenswrapper[4872]: I0126 09:11:09.979592 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:10 crc kubenswrapper[4872]: I0126 09:11:10.227409 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:11:10 crc kubenswrapper[4872]: I0126 09:11:10.227476 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:11:10 crc kubenswrapper[4872]: I0126 09:11:10.227520 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:11:10 crc kubenswrapper[4872]: I0126 09:11:10.228177 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd"} pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 09:11:10 crc kubenswrapper[4872]: I0126 09:11:10.228229 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" containerID="cri-o://28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd" gracePeriod=600 Jan 26 09:11:11 crc kubenswrapper[4872]: E0126 09:11:11.584093 4872 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:11 crc kubenswrapper[4872]: E0126 09:11:11.585030 4872 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:11 crc kubenswrapper[4872]: E0126 09:11:11.585334 4872 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:11 crc kubenswrapper[4872]: E0126 09:11:11.585586 4872 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:11 crc kubenswrapper[4872]: E0126 09:11:11.586056 4872 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:11 crc kubenswrapper[4872]: I0126 09:11:11.586221 4872 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Jan 26 09:11:11 crc kubenswrapper[4872]: E0126 09:11:11.586740 4872 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" interval="200ms" Jan 26 09:11:11 crc kubenswrapper[4872]: E0126 09:11:11.788324 4872 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" interval="400ms" Jan 26 09:11:12 crc kubenswrapper[4872]: E0126 09:11:12.017279 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:11:12Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:11:12Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:11:12Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-01-26T09:11:12Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:12 crc kubenswrapper[4872]: E0126 09:11:12.017544 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:12 crc kubenswrapper[4872]: E0126 09:11:12.017754 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:12 crc kubenswrapper[4872]: E0126 09:11:12.017974 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:12 crc kubenswrapper[4872]: E0126 09:11:12.018149 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:12 crc kubenswrapper[4872]: E0126 09:11:12.018170 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Jan 26 09:11:12 crc kubenswrapper[4872]: E0126 09:11:12.189256 4872 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" interval="800ms" Jan 26 09:11:12 crc kubenswrapper[4872]: I0126 09:11:12.955808 4872 generic.go:334] "Generic (PLEG): container finished" podID="fca44d96-a000-4bf2-8283-a937b0192880" containerID="28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd" exitCode=0 Jan 26 09:11:12 crc kubenswrapper[4872]: I0126 09:11:12.955855 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerDied","Data":"28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd"} Jan 26 09:11:12 crc kubenswrapper[4872]: E0126 09:11:12.990700 4872 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" interval="1.6s" Jan 26 09:11:13 crc kubenswrapper[4872]: I0126 09:11:13.972568 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerStarted","Data":"fc57e5384dec6c48143f4ceff8720fb743fb9d09433bb2bd3b7ba7d5cc00b27c"} Jan 26 09:11:13 crc kubenswrapper[4872]: I0126 09:11:13.974382 4872 status_manager.go:851] "Failed to get status for pod" podUID="41ca443e-2b21-4369-851c-30b312340df2" pod="openshift-marketplace/redhat-marketplace-rnmnc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rnmnc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:13 crc kubenswrapper[4872]: I0126 09:11:13.974615 4872 status_manager.go:851] "Failed to get status for pod" podUID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" pod="openshift-marketplace/redhat-operators-d5qrt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d5qrt\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:13 crc kubenswrapper[4872]: I0126 09:11:13.974785 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:13 crc kubenswrapper[4872]: I0126 09:11:13.974977 4872 status_manager.go:851] "Failed to get status for pod" podUID="5f25adda-7523-4fba-9f9e-32583f0d9ad6" pod="openshift-marketplace/certified-operators-5mk2s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5mk2s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:13 crc kubenswrapper[4872]: I0126 09:11:13.975147 4872 status_manager.go:851] "Failed to get status for pod" podUID="fca44d96-a000-4bf2-8283-a937b0192880" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-gt4gn\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:13 crc kubenswrapper[4872]: I0126 09:11:13.975352 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:13 crc kubenswrapper[4872]: I0126 09:11:13.975550 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:14 crc kubenswrapper[4872]: E0126 09:11:14.591511 4872 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.70:6443: connect: connection refused" interval="3.2s" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.183412 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.184241 4872 status_manager.go:851] "Failed to get status for pod" podUID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" pod="openshift-marketplace/redhat-operators-d5qrt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d5qrt\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.184765 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.185189 4872 status_manager.go:851] "Failed to get status for pod" podUID="5f25adda-7523-4fba-9f9e-32583f0d9ad6" pod="openshift-marketplace/certified-operators-5mk2s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5mk2s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.185545 4872 status_manager.go:851] "Failed to get status for pod" podUID="fca44d96-a000-4bf2-8283-a937b0192880" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-gt4gn\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.185830 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.186191 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.186689 4872 status_manager.go:851] "Failed to get status for pod" podUID="41ca443e-2b21-4369-851c-30b312340df2" pod="openshift-marketplace/redhat-marketplace-rnmnc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rnmnc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.199411 4872 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f24e9e28-d881-4922-9462-87bccec0c844" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.199447 4872 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f24e9e28-d881-4922-9462-87bccec0c844" Jan 26 09:11:15 crc kubenswrapper[4872]: E0126 09:11:15.199994 4872 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.200508 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:15 crc kubenswrapper[4872]: W0126 09:11:15.223255 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-a0c753fc181c46617e36d4bd6614334c18d274265d4869863323ba4f642f56b7 WatchSource:0}: Error finding container a0c753fc181c46617e36d4bd6614334c18d274265d4869863323ba4f642f56b7: Status 404 returned error can't find the container with id a0c753fc181c46617e36d4bd6614334c18d274265d4869863323ba4f642f56b7 Jan 26 09:11:15 crc kubenswrapper[4872]: E0126 09:11:15.225942 4872 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.70:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" volumeName="registry-storage" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.989058 4872 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="884591f91fb23b0a75ad6a7395fafc47e92ff7844e98b5f5482354d28856cb80" exitCode=0 Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.989332 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"884591f91fb23b0a75ad6a7395fafc47e92ff7844e98b5f5482354d28856cb80"} Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.989520 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a0c753fc181c46617e36d4bd6614334c18d274265d4869863323ba4f642f56b7"} Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.989876 4872 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f24e9e28-d881-4922-9462-87bccec0c844" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.989891 4872 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f24e9e28-d881-4922-9462-87bccec0c844" Jan 26 09:11:15 crc kubenswrapper[4872]: E0126 09:11:15.990427 4872 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.990437 4872 status_manager.go:851] "Failed to get status for pod" podUID="faac185b-bdff-4f6d-801a-1458123a20b0" pod="openshift-marketplace/community-operators-rjj7m" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-rjj7m\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.990889 4872 status_manager.go:851] "Failed to get status for pod" podUID="41ca443e-2b21-4369-851c-30b312340df2" pod="openshift-marketplace/redhat-marketplace-rnmnc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-rnmnc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.991263 4872 status_manager.go:851] "Failed to get status for pod" podUID="fccd3bc2-18ff-40d4-b862-7257f6cebc5b" pod="openshift-marketplace/redhat-operators-d5qrt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-d5qrt\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.991849 4872 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.992191 4872 status_manager.go:851] "Failed to get status for pod" podUID="5f25adda-7523-4fba-9f9e-32583f0d9ad6" pod="openshift-marketplace/certified-operators-5mk2s" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-5mk2s\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.992490 4872 status_manager.go:851] "Failed to get status for pod" podUID="fca44d96-a000-4bf2-8283-a937b0192880" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-gt4gn\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:15 crc kubenswrapper[4872]: I0126 09:11:15.992761 4872 status_manager.go:851] "Failed to get status for pod" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.70:6443: connect: connection refused" Jan 26 09:11:16 crc kubenswrapper[4872]: I0126 09:11:16.915603 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" podUID="596ad6ba-b5ec-471c-96b6-1b24d525b163" containerName="oauth-openshift" containerID="cri-o://a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf" gracePeriod=15 Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.007652 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"109ae6cad2f2a4a15eb6e19cf3a35a297b803a2feed1d8a4be164617e5e3ccf4"} Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.007706 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3cc090c78eeee734cab924e2fab61be0973481bf4d65e9cea07509b12106b30f"} Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.007719 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"758823745dea44cf901824fb89a6c81bfb82228e6fdfa008e0f762c0446f5619"} Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.435275 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.493850 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-trusted-ca-bundle\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.494054 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-idp-0-file-data\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.494139 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-session\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.494239 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-error\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.494349 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8tzq\" (UniqueName: \"kubernetes.io/projected/596ad6ba-b5ec-471c-96b6-1b24d525b163-kube-api-access-l8tzq\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.494469 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-ocp-branding-template\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.495740 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-login\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.495890 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/596ad6ba-b5ec-471c-96b6-1b24d525b163-audit-dir\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.495977 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-provider-selection\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.496063 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-router-certs\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.496164 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-service-ca\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.496262 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-serving-cert\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.496352 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-cliconfig\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.495998 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.496526 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-audit-policies\") pod \"596ad6ba-b5ec-471c-96b6-1b24d525b163\" (UID: \"596ad6ba-b5ec-471c-96b6-1b24d525b163\") " Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.496861 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.497339 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.498264 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.498334 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/596ad6ba-b5ec-471c-96b6-1b24d525b163-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.498891 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.505067 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.505990 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/596ad6ba-b5ec-471c-96b6-1b24d525b163-kube-api-access-l8tzq" (OuterVolumeSpecName: "kube-api-access-l8tzq") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "kube-api-access-l8tzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.506634 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.508438 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.508727 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.508747 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.509067 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.509308 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.513331 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "596ad6ba-b5ec-471c-96b6-1b24d525b163" (UID: "596ad6ba-b5ec-471c-96b6-1b24d525b163"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598267 4872 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-audit-policies\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598331 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598347 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598366 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598386 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8tzq\" (UniqueName: \"kubernetes.io/projected/596ad6ba-b5ec-471c-96b6-1b24d525b163-kube-api-access-l8tzq\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598398 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598409 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598420 4872 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/596ad6ba-b5ec-471c-96b6-1b24d525b163-audit-dir\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598431 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598443 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598453 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598464 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:17 crc kubenswrapper[4872]: I0126 09:11:17.598477 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/596ad6ba-b5ec-471c-96b6-1b24d525b163-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Jan 26 09:11:18 crc kubenswrapper[4872]: I0126 09:11:18.013662 4872 generic.go:334] "Generic (PLEG): container finished" podID="596ad6ba-b5ec-471c-96b6-1b24d525b163" containerID="a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf" exitCode=0 Jan 26 09:11:18 crc kubenswrapper[4872]: I0126 09:11:18.013771 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" Jan 26 09:11:18 crc kubenswrapper[4872]: I0126 09:11:18.014441 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" event={"ID":"596ad6ba-b5ec-471c-96b6-1b24d525b163","Type":"ContainerDied","Data":"a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf"} Jan 26 09:11:18 crc kubenswrapper[4872]: I0126 09:11:18.014475 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-mk4bd" event={"ID":"596ad6ba-b5ec-471c-96b6-1b24d525b163","Type":"ContainerDied","Data":"e5819feef382d300f684035d9b58e6865d004da941fb9e9c5974b810413cea94"} Jan 26 09:11:18 crc kubenswrapper[4872]: I0126 09:11:18.014493 4872 scope.go:117] "RemoveContainer" containerID="a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf" Jan 26 09:11:18 crc kubenswrapper[4872]: I0126 09:11:18.020580 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"10a292474de184a56e43ffb0c3bbc1c076a2feff2c20f0fdd7e47d340ee20fee"} Jan 26 09:11:18 crc kubenswrapper[4872]: I0126 09:11:18.020617 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f0a45bf796215feaf81a6337a9ab5a086e2cc3a4e2e5b1e69f41d880bf0de8b3"} Jan 26 09:11:18 crc kubenswrapper[4872]: I0126 09:11:18.020968 4872 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f24e9e28-d881-4922-9462-87bccec0c844" Jan 26 09:11:18 crc kubenswrapper[4872]: I0126 09:11:18.021011 4872 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f24e9e28-d881-4922-9462-87bccec0c844" Jan 26 09:11:18 crc kubenswrapper[4872]: I0126 09:11:18.021256 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:18 crc kubenswrapper[4872]: I0126 09:11:18.034844 4872 scope.go:117] "RemoveContainer" containerID="a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf" Jan 26 09:11:18 crc kubenswrapper[4872]: E0126 09:11:18.036357 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf\": container with ID starting with a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf not found: ID does not exist" containerID="a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf" Jan 26 09:11:18 crc kubenswrapper[4872]: I0126 09:11:18.036405 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf"} err="failed to get container status \"a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf\": rpc error: code = NotFound desc = could not find container \"a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf\": container with ID starting with a5aeaa50a21ad09320370493d3ff110c7bcc4d1877bbc01ded83d61b011a4eaf not found: ID does not exist" Jan 26 09:11:19 crc kubenswrapper[4872]: I0126 09:11:19.035885 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 26 09:11:19 crc kubenswrapper[4872]: I0126 09:11:19.035942 4872 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96" exitCode=1 Jan 26 09:11:19 crc kubenswrapper[4872]: I0126 09:11:19.035979 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96"} Jan 26 09:11:19 crc kubenswrapper[4872]: I0126 09:11:19.036513 4872 scope.go:117] "RemoveContainer" containerID="291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96" Jan 26 09:11:19 crc kubenswrapper[4872]: I0126 09:11:19.666247 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 09:11:20 crc kubenswrapper[4872]: I0126 09:11:20.046100 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 26 09:11:20 crc kubenswrapper[4872]: I0126 09:11:20.046463 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3ee2424d97017d5db918279fd9df2cbbb8e1a12f34930e19474e9141e3080ca7"} Jan 26 09:11:20 crc kubenswrapper[4872]: I0126 09:11:20.201629 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:20 crc kubenswrapper[4872]: I0126 09:11:20.202029 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:20 crc kubenswrapper[4872]: I0126 09:11:20.207313 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:20 crc kubenswrapper[4872]: I0126 09:11:20.787660 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 09:11:20 crc kubenswrapper[4872]: I0126 09:11:20.787928 4872 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 26 09:11:20 crc kubenswrapper[4872]: I0126 09:11:20.788009 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 26 09:11:23 crc kubenswrapper[4872]: I0126 09:11:23.037305 4872 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:23 crc kubenswrapper[4872]: I0126 09:11:23.063714 4872 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f24e9e28-d881-4922-9462-87bccec0c844" Jan 26 09:11:23 crc kubenswrapper[4872]: I0126 09:11:23.064099 4872 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f24e9e28-d881-4922-9462-87bccec0c844" Jan 26 09:11:23 crc kubenswrapper[4872]: I0126 09:11:23.069112 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:23 crc kubenswrapper[4872]: I0126 09:11:23.149312 4872 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="9c23b259-6d55-4cc3-a12b-304073f9f0df" Jan 26 09:11:24 crc kubenswrapper[4872]: I0126 09:11:24.071252 4872 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f24e9e28-d881-4922-9462-87bccec0c844" Jan 26 09:11:24 crc kubenswrapper[4872]: I0126 09:11:24.071568 4872 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f24e9e28-d881-4922-9462-87bccec0c844" Jan 26 09:11:27 crc kubenswrapper[4872]: I0126 09:11:27.208546 4872 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="9c23b259-6d55-4cc3-a12b-304073f9f0df" Jan 26 09:11:29 crc kubenswrapper[4872]: I0126 09:11:29.370329 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Jan 26 09:11:29 crc kubenswrapper[4872]: I0126 09:11:29.653358 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Jan 26 09:11:29 crc kubenswrapper[4872]: I0126 09:11:29.665312 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 09:11:29 crc kubenswrapper[4872]: I0126 09:11:29.835878 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Jan 26 09:11:30 crc kubenswrapper[4872]: I0126 09:11:30.345246 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Jan 26 09:11:30 crc kubenswrapper[4872]: I0126 09:11:30.452558 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Jan 26 09:11:30 crc kubenswrapper[4872]: I0126 09:11:30.710603 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Jan 26 09:11:30 crc kubenswrapper[4872]: I0126 09:11:30.736369 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Jan 26 09:11:30 crc kubenswrapper[4872]: I0126 09:11:30.741273 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Jan 26 09:11:30 crc kubenswrapper[4872]: I0126 09:11:30.788038 4872 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 26 09:11:30 crc kubenswrapper[4872]: I0126 09:11:30.788135 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 26 09:11:30 crc kubenswrapper[4872]: I0126 09:11:30.954233 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Jan 26 09:11:31 crc kubenswrapper[4872]: I0126 09:11:31.179669 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Jan 26 09:11:31 crc kubenswrapper[4872]: I0126 09:11:31.197367 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Jan 26 09:11:31 crc kubenswrapper[4872]: I0126 09:11:31.198035 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Jan 26 09:11:31 crc kubenswrapper[4872]: I0126 09:11:31.548322 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Jan 26 09:11:31 crc kubenswrapper[4872]: I0126 09:11:31.781949 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Jan 26 09:11:31 crc kubenswrapper[4872]: I0126 09:11:31.870645 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Jan 26 09:11:31 crc kubenswrapper[4872]: I0126 09:11:31.871324 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Jan 26 09:11:32 crc kubenswrapper[4872]: I0126 09:11:32.085208 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Jan 26 09:11:32 crc kubenswrapper[4872]: I0126 09:11:32.130567 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Jan 26 09:11:32 crc kubenswrapper[4872]: I0126 09:11:32.254881 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Jan 26 09:11:32 crc kubenswrapper[4872]: I0126 09:11:32.300716 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Jan 26 09:11:32 crc kubenswrapper[4872]: I0126 09:11:32.457685 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Jan 26 09:11:32 crc kubenswrapper[4872]: I0126 09:11:32.499250 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Jan 26 09:11:32 crc kubenswrapper[4872]: I0126 09:11:32.878852 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Jan 26 09:11:33 crc kubenswrapper[4872]: I0126 09:11:33.101184 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Jan 26 09:11:33 crc kubenswrapper[4872]: I0126 09:11:33.631918 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Jan 26 09:11:33 crc kubenswrapper[4872]: I0126 09:11:33.930259 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Jan 26 09:11:34 crc kubenswrapper[4872]: I0126 09:11:34.063791 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Jan 26 09:11:34 crc kubenswrapper[4872]: I0126 09:11:34.598616 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Jan 26 09:11:34 crc kubenswrapper[4872]: I0126 09:11:34.629094 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Jan 26 09:11:34 crc kubenswrapper[4872]: I0126 09:11:34.639194 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 26 09:11:34 crc kubenswrapper[4872]: I0126 09:11:34.691623 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Jan 26 09:11:34 crc kubenswrapper[4872]: I0126 09:11:34.966972 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Jan 26 09:11:34 crc kubenswrapper[4872]: I0126 09:11:34.984821 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Jan 26 09:11:35 crc kubenswrapper[4872]: I0126 09:11:35.073868 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Jan 26 09:11:35 crc kubenswrapper[4872]: I0126 09:11:35.132377 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Jan 26 09:11:35 crc kubenswrapper[4872]: I0126 09:11:35.251661 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Jan 26 09:11:35 crc kubenswrapper[4872]: I0126 09:11:35.533357 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Jan 26 09:11:35 crc kubenswrapper[4872]: I0126 09:11:35.536016 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 26 09:11:35 crc kubenswrapper[4872]: I0126 09:11:35.631301 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 26 09:11:35 crc kubenswrapper[4872]: I0126 09:11:35.655718 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Jan 26 09:11:35 crc kubenswrapper[4872]: I0126 09:11:35.854263 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Jan 26 09:11:35 crc kubenswrapper[4872]: I0126 09:11:35.893102 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Jan 26 09:11:36 crc kubenswrapper[4872]: I0126 09:11:36.055646 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Jan 26 09:11:36 crc kubenswrapper[4872]: I0126 09:11:36.122341 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Jan 26 09:11:36 crc kubenswrapper[4872]: I0126 09:11:36.692755 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Jan 26 09:11:36 crc kubenswrapper[4872]: I0126 09:11:36.809742 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Jan 26 09:11:36 crc kubenswrapper[4872]: I0126 09:11:36.823669 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Jan 26 09:11:37 crc kubenswrapper[4872]: I0126 09:11:37.036102 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Jan 26 09:11:37 crc kubenswrapper[4872]: I0126 09:11:37.317389 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 26 09:11:37 crc kubenswrapper[4872]: I0126 09:11:37.387658 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 26 09:11:37 crc kubenswrapper[4872]: I0126 09:11:37.605762 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Jan 26 09:11:37 crc kubenswrapper[4872]: I0126 09:11:37.633495 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Jan 26 09:11:37 crc kubenswrapper[4872]: I0126 09:11:37.699708 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Jan 26 09:11:37 crc kubenswrapper[4872]: I0126 09:11:37.866783 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Jan 26 09:11:37 crc kubenswrapper[4872]: I0126 09:11:37.890039 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Jan 26 09:11:37 crc kubenswrapper[4872]: I0126 09:11:37.925991 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Jan 26 09:11:37 crc kubenswrapper[4872]: I0126 09:11:37.971538 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Jan 26 09:11:38 crc kubenswrapper[4872]: I0126 09:11:38.226383 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Jan 26 09:11:38 crc kubenswrapper[4872]: I0126 09:11:38.870072 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Jan 26 09:11:38 crc kubenswrapper[4872]: I0126 09:11:38.955421 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 26 09:11:38 crc kubenswrapper[4872]: I0126 09:11:38.972914 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Jan 26 09:11:39 crc kubenswrapper[4872]: I0126 09:11:39.128911 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Jan 26 09:11:39 crc kubenswrapper[4872]: I0126 09:11:39.155301 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Jan 26 09:11:39 crc kubenswrapper[4872]: I0126 09:11:39.157565 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Jan 26 09:11:39 crc kubenswrapper[4872]: I0126 09:11:39.266227 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Jan 26 09:11:39 crc kubenswrapper[4872]: I0126 09:11:39.310712 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Jan 26 09:11:39 crc kubenswrapper[4872]: I0126 09:11:39.359387 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Jan 26 09:11:39 crc kubenswrapper[4872]: I0126 09:11:39.597063 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Jan 26 09:11:39 crc kubenswrapper[4872]: I0126 09:11:39.614212 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Jan 26 09:11:39 crc kubenswrapper[4872]: I0126 09:11:39.825541 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Jan 26 09:11:39 crc kubenswrapper[4872]: I0126 09:11:39.941983 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Jan 26 09:11:39 crc kubenswrapper[4872]: I0126 09:11:39.973894 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.009158 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.097866 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.136308 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.336000 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.370597 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.453091 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.463603 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.478560 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.632245 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.672793 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.723922 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.788445 4872 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.788522 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.788598 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.789333 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"3ee2424d97017d5db918279fd9df2cbbb8e1a12f34930e19474e9141e3080ca7"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.789440 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://3ee2424d97017d5db918279fd9df2cbbb8e1a12f34930e19474e9141e3080ca7" gracePeriod=30 Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.822260 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Jan 26 09:11:40 crc kubenswrapper[4872]: I0126 09:11:40.890821 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Jan 26 09:11:41 crc kubenswrapper[4872]: I0126 09:11:41.318759 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 26 09:11:41 crc kubenswrapper[4872]: I0126 09:11:41.409336 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Jan 26 09:11:41 crc kubenswrapper[4872]: I0126 09:11:41.462253 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Jan 26 09:11:41 crc kubenswrapper[4872]: I0126 09:11:41.524449 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Jan 26 09:11:41 crc kubenswrapper[4872]: I0126 09:11:41.714089 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Jan 26 09:11:41 crc kubenswrapper[4872]: I0126 09:11:41.766380 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Jan 26 09:11:41 crc kubenswrapper[4872]: I0126 09:11:41.773356 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.000895 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.030724 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.074762 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.102152 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.189088 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.262728 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.308720 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.352497 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.507157 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.547707 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.660420 4872 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.665099 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.684631 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.716553 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.733445 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.778448 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.800468 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.805487 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.846285 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.949113 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Jan 26 09:11:42 crc kubenswrapper[4872]: I0126 09:11:42.949113 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.009570 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.047956 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.134223 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.229650 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.276485 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.352344 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.383899 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.438018 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.518523 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.527232 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.568568 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.674209 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.708508 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.726448 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.766551 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.911064 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Jan 26 09:11:43 crc kubenswrapper[4872]: I0126 09:11:43.966773 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.018469 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.021585 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.117215 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.191874 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.215492 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.258050 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.364919 4872 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.504109 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.558041 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.599616 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.652536 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.663688 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.671085 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.725123 4872 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.728437 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Jan 26 09:11:44 crc kubenswrapper[4872]: I0126 09:11:44.859322 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.033634 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.144574 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.149476 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.200740 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.201196 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.247593 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.273344 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.290446 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.406306 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.448102 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.455268 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.500751 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.501859 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.514418 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.533644 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.545166 4872 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.558991 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.649102 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.773521 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.827281 4872 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.828619 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=41.828602056 podStartE2EDuration="41.828602056s" podCreationTimestamp="2026-01-26 09:11:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:11:23.051614861 +0000 UTC m=+216.360454672" watchObservedRunningTime="2026-01-26 09:11:45.828602056 +0000 UTC m=+239.137441867" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.833089 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-mk4bd","openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.833154 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2","openshift-kube-apiserver/kube-apiserver-crc"] Jan 26 09:11:45 crc kubenswrapper[4872]: E0126 09:11:45.833438 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" containerName="installer" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.833464 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" containerName="installer" Jan 26 09:11:45 crc kubenswrapper[4872]: E0126 09:11:45.833478 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="596ad6ba-b5ec-471c-96b6-1b24d525b163" containerName="oauth-openshift" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.833488 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="596ad6ba-b5ec-471c-96b6-1b24d525b163" containerName="oauth-openshift" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.833649 4872 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f24e9e28-d881-4922-9462-87bccec0c844" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.833688 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="717a9b10-82c6-4474-8715-0b34ff4e391e" containerName="installer" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.833711 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="596ad6ba-b5ec-471c-96b6-1b24d525b163" containerName="oauth-openshift" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.833694 4872 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f24e9e28-d881-4922-9462-87bccec0c844" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.834936 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.840326 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.840364 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.840489 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.840966 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.844788 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.844927 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.845191 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.845310 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.845350 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.845394 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.845451 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.845539 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.845645 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.849215 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.856162 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.862093 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.895451 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.895427403 podStartE2EDuration="22.895427403s" podCreationTimestamp="2026-01-26 09:11:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:11:45.891121334 +0000 UTC m=+239.199961135" watchObservedRunningTime="2026-01-26 09:11:45.895427403 +0000 UTC m=+239.204267214" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.921674 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e2e65972-550c-44b6-9ea7-e1c574b55afc-audit-dir\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.921732 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-router-certs\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.921766 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.921792 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-user-template-login\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.921895 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.921926 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-serving-cert\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.921959 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-service-ca\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.921975 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.922002 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-session\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.922018 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-cliconfig\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.922034 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e2e65972-550c-44b6-9ea7-e1c574b55afc-audit-policies\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.922063 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.922079 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-user-template-error\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:45 crc kubenswrapper[4872]: I0126 09:11:45.922110 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxdrp\" (UniqueName: \"kubernetes.io/projected/e2e65972-550c-44b6-9ea7-e1c574b55afc-kube-api-access-zxdrp\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.003557 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.023722 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-router-certs\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.023790 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.023932 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-user-template-login\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.023951 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.023970 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-serving-cert\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.023994 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-service-ca\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.024011 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.024039 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-session\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.024056 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-cliconfig\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.024074 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e2e65972-550c-44b6-9ea7-e1c574b55afc-audit-policies\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.024104 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.024122 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-user-template-error\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.024153 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxdrp\" (UniqueName: \"kubernetes.io/projected/e2e65972-550c-44b6-9ea7-e1c574b55afc-kube-api-access-zxdrp\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.024177 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e2e65972-550c-44b6-9ea7-e1c574b55afc-audit-dir\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.024243 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e2e65972-550c-44b6-9ea7-e1c574b55afc-audit-dir\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.025571 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e2e65972-550c-44b6-9ea7-e1c574b55afc-audit-policies\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.025688 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-cliconfig\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.025729 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.026241 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-service-ca\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.030408 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-user-template-error\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.030943 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-router-certs\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.031894 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.031948 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-user-template-login\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.032291 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.035982 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-serving-cert\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.039281 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.039385 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e2e65972-550c-44b6-9ea7-e1c574b55afc-v4-0-config-system-session\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.046627 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxdrp\" (UniqueName: \"kubernetes.io/projected/e2e65972-550c-44b6-9ea7-e1c574b55afc-kube-api-access-zxdrp\") pod \"oauth-openshift-56f4fc5f47-s9nx2\" (UID: \"e2e65972-550c-44b6-9ea7-e1c574b55afc\") " pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.160453 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.297960 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.472949 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.571073 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.571943 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.594154 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.601996 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2"] Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.617340 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.833213 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.838868 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.840336 4872 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.908574 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.923813 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.936390 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.937362 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.937987 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.981005 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Jan 26 09:11:46 crc kubenswrapper[4872]: I0126 09:11:46.992057 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.036299 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.063962 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.093587 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.104332 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.111554 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.186785 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.192002 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="596ad6ba-b5ec-471c-96b6-1b24d525b163" path="/var/lib/kubelet/pods/596ad6ba-b5ec-471c-96b6-1b24d525b163/volumes" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.201964 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.212639 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.213992 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" event={"ID":"e2e65972-550c-44b6-9ea7-e1c574b55afc","Type":"ContainerStarted","Data":"74601d023729d80ea707140a7dcd909668f775d092e31500c150261fd93f35dd"} Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.214052 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" event={"ID":"e2e65972-550c-44b6-9ea7-e1c574b55afc","Type":"ContainerStarted","Data":"57f1c00f38d0714c5539ef206cd7ca0b54a10edd9b54ac127c8a6e5720f23dc8"} Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.214776 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.238535 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" podStartSLOduration=56.238513163 podStartE2EDuration="56.238513163s" podCreationTimestamp="2026-01-26 09:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:11:47.236901199 +0000 UTC m=+240.545741000" watchObservedRunningTime="2026-01-26 09:11:47.238513163 +0000 UTC m=+240.547352964" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.297198 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-56f4fc5f47-s9nx2" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.321130 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.431828 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.486029 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.518910 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.571136 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.571225 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.580284 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.592420 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.701630 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.706955 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.818037 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.853308 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.869729 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.896111 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.946413 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Jan 26 09:11:47 crc kubenswrapper[4872]: I0126 09:11:47.988481 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Jan 26 09:11:48 crc kubenswrapper[4872]: I0126 09:11:48.110286 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Jan 26 09:11:48 crc kubenswrapper[4872]: I0126 09:11:48.220784 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Jan 26 09:11:48 crc kubenswrapper[4872]: I0126 09:11:48.344601 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Jan 26 09:11:48 crc kubenswrapper[4872]: I0126 09:11:48.488505 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Jan 26 09:11:48 crc kubenswrapper[4872]: I0126 09:11:48.812919 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Jan 26 09:11:48 crc kubenswrapper[4872]: I0126 09:11:48.944419 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Jan 26 09:11:49 crc kubenswrapper[4872]: I0126 09:11:49.289582 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Jan 26 09:11:49 crc kubenswrapper[4872]: I0126 09:11:49.320382 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Jan 26 09:11:49 crc kubenswrapper[4872]: I0126 09:11:49.525875 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Jan 26 09:11:49 crc kubenswrapper[4872]: I0126 09:11:49.528684 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Jan 26 09:11:49 crc kubenswrapper[4872]: I0126 09:11:49.708641 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Jan 26 09:11:49 crc kubenswrapper[4872]: I0126 09:11:49.718665 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Jan 26 09:11:50 crc kubenswrapper[4872]: I0126 09:11:50.072779 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Jan 26 09:11:50 crc kubenswrapper[4872]: I0126 09:11:50.194296 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Jan 26 09:11:50 crc kubenswrapper[4872]: I0126 09:11:50.365288 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Jan 26 09:11:50 crc kubenswrapper[4872]: I0126 09:11:50.375519 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Jan 26 09:11:50 crc kubenswrapper[4872]: I0126 09:11:50.531788 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 26 09:11:50 crc kubenswrapper[4872]: I0126 09:11:50.583668 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Jan 26 09:11:50 crc kubenswrapper[4872]: I0126 09:11:50.657863 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Jan 26 09:11:50 crc kubenswrapper[4872]: I0126 09:11:50.752169 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Jan 26 09:11:50 crc kubenswrapper[4872]: I0126 09:11:50.909630 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Jan 26 09:11:50 crc kubenswrapper[4872]: I0126 09:11:50.956145 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Jan 26 09:11:50 crc kubenswrapper[4872]: I0126 09:11:50.977362 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 26 09:11:51 crc kubenswrapper[4872]: I0126 09:11:51.016132 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Jan 26 09:11:51 crc kubenswrapper[4872]: I0126 09:11:51.107009 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Jan 26 09:11:51 crc kubenswrapper[4872]: I0126 09:11:51.245102 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 26 09:11:51 crc kubenswrapper[4872]: I0126 09:11:51.407582 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Jan 26 09:11:51 crc kubenswrapper[4872]: I0126 09:11:51.518976 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Jan 26 09:11:51 crc kubenswrapper[4872]: I0126 09:11:51.661147 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Jan 26 09:11:53 crc kubenswrapper[4872]: I0126 09:11:53.240097 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Jan 26 09:11:57 crc kubenswrapper[4872]: I0126 09:11:57.098222 4872 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 09:11:57 crc kubenswrapper[4872]: I0126 09:11:57.099728 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://66126fdf89160b3f4d8fee22a92d4e517fc226b6a05cbae1286525e4e07373af" gracePeriod=5 Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.314639 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.315523 4872 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="66126fdf89160b3f4d8fee22a92d4e517fc226b6a05cbae1286525e4e07373af" exitCode=137 Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.681603 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.681694 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.792941 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.793067 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.793135 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.793290 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.793327 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.793396 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.793456 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.793510 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.793605 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.793835 4872 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.793849 4872 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.793859 4872 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.793869 4872 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.804215 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:12:02 crc kubenswrapper[4872]: I0126 09:12:02.895771 4872 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:03 crc kubenswrapper[4872]: I0126 09:12:03.200260 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Jan 26 09:12:03 crc kubenswrapper[4872]: I0126 09:12:03.200889 4872 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Jan 26 09:12:03 crc kubenswrapper[4872]: I0126 09:12:03.221061 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 09:12:03 crc kubenswrapper[4872]: I0126 09:12:03.221188 4872 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="d6ef059b-bcd7-40bc-b28f-d1198b34e678" Jan 26 09:12:03 crc kubenswrapper[4872]: I0126 09:12:03.231155 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Jan 26 09:12:03 crc kubenswrapper[4872]: I0126 09:12:03.231204 4872 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="d6ef059b-bcd7-40bc-b28f-d1198b34e678" Jan 26 09:12:03 crc kubenswrapper[4872]: I0126 09:12:03.324341 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Jan 26 09:12:03 crc kubenswrapper[4872]: I0126 09:12:03.324433 4872 scope.go:117] "RemoveContainer" containerID="66126fdf89160b3f4d8fee22a92d4e517fc226b6a05cbae1286525e4e07373af" Jan 26 09:12:03 crc kubenswrapper[4872]: I0126 09:12:03.324544 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Jan 26 09:12:11 crc kubenswrapper[4872]: I0126 09:12:11.392985 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 26 09:12:11 crc kubenswrapper[4872]: I0126 09:12:11.397352 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Jan 26 09:12:11 crc kubenswrapper[4872]: I0126 09:12:11.397471 4872 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="3ee2424d97017d5db918279fd9df2cbbb8e1a12f34930e19474e9141e3080ca7" exitCode=137 Jan 26 09:12:11 crc kubenswrapper[4872]: I0126 09:12:11.397507 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"3ee2424d97017d5db918279fd9df2cbbb8e1a12f34930e19474e9141e3080ca7"} Jan 26 09:12:11 crc kubenswrapper[4872]: I0126 09:12:11.397542 4872 scope.go:117] "RemoveContainer" containerID="291d32c3d97b7eb377c185f1f08456a3d4ee9b3ed6ed61b9c0a51aff1fcf5b96" Jan 26 09:12:12 crc kubenswrapper[4872]: I0126 09:12:12.431138 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Jan 26 09:12:12 crc kubenswrapper[4872]: I0126 09:12:12.434168 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cf76e65ddb37ce3e8058a6563464701623647f0b26debcef3c05e7d7dc4db041"} Jan 26 09:12:19 crc kubenswrapper[4872]: I0126 09:12:19.666198 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 09:12:20 crc kubenswrapper[4872]: I0126 09:12:20.787748 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 09:12:20 crc kubenswrapper[4872]: I0126 09:12:20.795079 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 09:12:21 crc kubenswrapper[4872]: I0126 09:12:21.493306 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Jan 26 09:12:29 crc kubenswrapper[4872]: I0126 09:12:29.404395 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w"] Jan 26 09:12:29 crc kubenswrapper[4872]: I0126 09:12:29.405344 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" podUID="47ab1f7e-e14b-41e3-93b1-6c1b038f93a8" containerName="route-controller-manager" containerID="cri-o://636065f9e0fb91ddd2e5023f8178193560c135de1cd75199d6ca76c7be759d58" gracePeriod=30 Jan 26 09:12:29 crc kubenswrapper[4872]: I0126 09:12:29.412786 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kxnj5"] Jan 26 09:12:29 crc kubenswrapper[4872]: I0126 09:12:29.413210 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" podUID="3a3df101-99e3-41a0-a235-0876a7249434" containerName="controller-manager" containerID="cri-o://c2f1f4116f3dbcafa5f89510171a630b400c108e120639da2f9e2fb3db9052aa" gracePeriod=30 Jan 26 09:12:29 crc kubenswrapper[4872]: I0126 09:12:29.553093 4872 generic.go:334] "Generic (PLEG): container finished" podID="47ab1f7e-e14b-41e3-93b1-6c1b038f93a8" containerID="636065f9e0fb91ddd2e5023f8178193560c135de1cd75199d6ca76c7be759d58" exitCode=0 Jan 26 09:12:29 crc kubenswrapper[4872]: I0126 09:12:29.553591 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" event={"ID":"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8","Type":"ContainerDied","Data":"636065f9e0fb91ddd2e5023f8178193560c135de1cd75199d6ca76c7be759d58"} Jan 26 09:12:29 crc kubenswrapper[4872]: I0126 09:12:29.554925 4872 generic.go:334] "Generic (PLEG): container finished" podID="3a3df101-99e3-41a0-a235-0876a7249434" containerID="c2f1f4116f3dbcafa5f89510171a630b400c108e120639da2f9e2fb3db9052aa" exitCode=0 Jan 26 09:12:29 crc kubenswrapper[4872]: I0126 09:12:29.554951 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" event={"ID":"3a3df101-99e3-41a0-a235-0876a7249434","Type":"ContainerDied","Data":"c2f1f4116f3dbcafa5f89510171a630b400c108e120639da2f9e2fb3db9052aa"} Jan 26 09:12:29 crc kubenswrapper[4872]: I0126 09:12:29.882244 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:12:29 crc kubenswrapper[4872]: I0126 09:12:29.932764 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.064093 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-client-ca\") pod \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.064171 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-proxy-ca-bundles\") pod \"3a3df101-99e3-41a0-a235-0876a7249434\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.064214 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2cpv\" (UniqueName: \"kubernetes.io/projected/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-kube-api-access-g2cpv\") pod \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.064241 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-serving-cert\") pod \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.064275 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz9xr\" (UniqueName: \"kubernetes.io/projected/3a3df101-99e3-41a0-a235-0876a7249434-kube-api-access-zz9xr\") pod \"3a3df101-99e3-41a0-a235-0876a7249434\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.064305 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-client-ca\") pod \"3a3df101-99e3-41a0-a235-0876a7249434\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.064329 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-config\") pod \"3a3df101-99e3-41a0-a235-0876a7249434\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.064350 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3df101-99e3-41a0-a235-0876a7249434-serving-cert\") pod \"3a3df101-99e3-41a0-a235-0876a7249434\" (UID: \"3a3df101-99e3-41a0-a235-0876a7249434\") " Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.064403 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-config\") pod \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\" (UID: \"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8\") " Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.065329 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3a3df101-99e3-41a0-a235-0876a7249434" (UID: "3a3df101-99e3-41a0-a235-0876a7249434"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.065455 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-config" (OuterVolumeSpecName: "config") pod "3a3df101-99e3-41a0-a235-0876a7249434" (UID: "3a3df101-99e3-41a0-a235-0876a7249434"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.065626 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-client-ca" (OuterVolumeSpecName: "client-ca") pod "3a3df101-99e3-41a0-a235-0876a7249434" (UID: "3a3df101-99e3-41a0-a235-0876a7249434"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.067685 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-client-ca" (OuterVolumeSpecName: "client-ca") pod "47ab1f7e-e14b-41e3-93b1-6c1b038f93a8" (UID: "47ab1f7e-e14b-41e3-93b1-6c1b038f93a8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.070505 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-config" (OuterVolumeSpecName: "config") pod "47ab1f7e-e14b-41e3-93b1-6c1b038f93a8" (UID: "47ab1f7e-e14b-41e3-93b1-6c1b038f93a8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.073113 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a3df101-99e3-41a0-a235-0876a7249434-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3a3df101-99e3-41a0-a235-0876a7249434" (UID: "3a3df101-99e3-41a0-a235-0876a7249434"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.074521 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a3df101-99e3-41a0-a235-0876a7249434-kube-api-access-zz9xr" (OuterVolumeSpecName: "kube-api-access-zz9xr") pod "3a3df101-99e3-41a0-a235-0876a7249434" (UID: "3a3df101-99e3-41a0-a235-0876a7249434"). InnerVolumeSpecName "kube-api-access-zz9xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.075848 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "47ab1f7e-e14b-41e3-93b1-6c1b038f93a8" (UID: "47ab1f7e-e14b-41e3-93b1-6c1b038f93a8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.076100 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-kube-api-access-g2cpv" (OuterVolumeSpecName: "kube-api-access-g2cpv") pod "47ab1f7e-e14b-41e3-93b1-6c1b038f93a8" (UID: "47ab1f7e-e14b-41e3-93b1-6c1b038f93a8"). InnerVolumeSpecName "kube-api-access-g2cpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.165921 4872 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.165977 4872 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.165999 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2cpv\" (UniqueName: \"kubernetes.io/projected/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-kube-api-access-g2cpv\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.166019 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.166036 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz9xr\" (UniqueName: \"kubernetes.io/projected/3a3df101-99e3-41a0-a235-0876a7249434-kube-api-access-zz9xr\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.166052 4872 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-client-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.166070 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3a3df101-99e3-41a0-a235-0876a7249434-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.166084 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3a3df101-99e3-41a0-a235-0876a7249434-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.166099 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.562287 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" event={"ID":"47ab1f7e-e14b-41e3-93b1-6c1b038f93a8","Type":"ContainerDied","Data":"ec8a4fe827f1a2aba42ae3f47a99412d92e051f12fd02c16644e68273fc5cb7d"} Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.562362 4872 scope.go:117] "RemoveContainer" containerID="636065f9e0fb91ddd2e5023f8178193560c135de1cd75199d6ca76c7be759d58" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.562505 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.574129 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" event={"ID":"3a3df101-99e3-41a0-a235-0876a7249434","Type":"ContainerDied","Data":"82ce8d27ada799de6b5b637d607ddc613ea37ca80e7163380080dd9d7a7006a6"} Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.574219 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-kxnj5" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.591654 4872 scope.go:117] "RemoveContainer" containerID="c2f1f4116f3dbcafa5f89510171a630b400c108e120639da2f9e2fb3db9052aa" Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.618625 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kxnj5"] Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.622355 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-kxnj5"] Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.633352 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w"] Jan 26 09:12:30 crc kubenswrapper[4872]: I0126 09:12:30.637503 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pzn5w"] Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.194230 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a3df101-99e3-41a0-a235-0876a7249434" path="/var/lib/kubelet/pods/3a3df101-99e3-41a0-a235-0876a7249434/volumes" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.194764 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47ab1f7e-e14b-41e3-93b1-6c1b038f93a8" path="/var/lib/kubelet/pods/47ab1f7e-e14b-41e3-93b1-6c1b038f93a8/volumes" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.314766 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-57b6756bff-vmjcv"] Jan 26 09:12:31 crc kubenswrapper[4872]: E0126 09:12:31.315271 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a3df101-99e3-41a0-a235-0876a7249434" containerName="controller-manager" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.315312 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a3df101-99e3-41a0-a235-0876a7249434" containerName="controller-manager" Jan 26 09:12:31 crc kubenswrapper[4872]: E0126 09:12:31.315344 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ab1f7e-e14b-41e3-93b1-6c1b038f93a8" containerName="route-controller-manager" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.315357 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ab1f7e-e14b-41e3-93b1-6c1b038f93a8" containerName="route-controller-manager" Jan 26 09:12:31 crc kubenswrapper[4872]: E0126 09:12:31.315401 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.315416 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.315616 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ab1f7e-e14b-41e3-93b1-6c1b038f93a8" containerName="route-controller-manager" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.315662 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.315685 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a3df101-99e3-41a0-a235-0876a7249434" containerName="controller-manager" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.316442 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.318642 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.319729 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.320771 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.320773 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj"] Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.321247 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.321549 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.321901 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.321998 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.324941 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.325245 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.325581 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.326053 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.326222 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.326378 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.332189 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57b6756bff-vmjcv"] Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.337451 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.353596 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj"] Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.485337 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92fe1566-b184-4337-b3f6-15f1abd8b0b1-config\") pod \"route-controller-manager-67c6b664cd-bqjgj\" (UID: \"92fe1566-b184-4337-b3f6-15f1abd8b0b1\") " pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.485402 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5239d54f-7760-4e80-8f46-850a8ce75147-client-ca\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.485433 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfrmp\" (UniqueName: \"kubernetes.io/projected/92fe1566-b184-4337-b3f6-15f1abd8b0b1-kube-api-access-mfrmp\") pod \"route-controller-manager-67c6b664cd-bqjgj\" (UID: \"92fe1566-b184-4337-b3f6-15f1abd8b0b1\") " pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.485452 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5239d54f-7760-4e80-8f46-850a8ce75147-config\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.485470 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92fe1566-b184-4337-b3f6-15f1abd8b0b1-client-ca\") pod \"route-controller-manager-67c6b664cd-bqjgj\" (UID: \"92fe1566-b184-4337-b3f6-15f1abd8b0b1\") " pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.485492 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxvkp\" (UniqueName: \"kubernetes.io/projected/5239d54f-7760-4e80-8f46-850a8ce75147-kube-api-access-bxvkp\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.485529 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5239d54f-7760-4e80-8f46-850a8ce75147-serving-cert\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.485543 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5239d54f-7760-4e80-8f46-850a8ce75147-proxy-ca-bundles\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.485575 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92fe1566-b184-4337-b3f6-15f1abd8b0b1-serving-cert\") pod \"route-controller-manager-67c6b664cd-bqjgj\" (UID: \"92fe1566-b184-4337-b3f6-15f1abd8b0b1\") " pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.589096 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5239d54f-7760-4e80-8f46-850a8ce75147-serving-cert\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.589185 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5239d54f-7760-4e80-8f46-850a8ce75147-proxy-ca-bundles\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.590083 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92fe1566-b184-4337-b3f6-15f1abd8b0b1-serving-cert\") pod \"route-controller-manager-67c6b664cd-bqjgj\" (UID: \"92fe1566-b184-4337-b3f6-15f1abd8b0b1\") " pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.590153 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92fe1566-b184-4337-b3f6-15f1abd8b0b1-config\") pod \"route-controller-manager-67c6b664cd-bqjgj\" (UID: \"92fe1566-b184-4337-b3f6-15f1abd8b0b1\") " pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.590204 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5239d54f-7760-4e80-8f46-850a8ce75147-client-ca\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.590245 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfrmp\" (UniqueName: \"kubernetes.io/projected/92fe1566-b184-4337-b3f6-15f1abd8b0b1-kube-api-access-mfrmp\") pod \"route-controller-manager-67c6b664cd-bqjgj\" (UID: \"92fe1566-b184-4337-b3f6-15f1abd8b0b1\") " pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.590289 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5239d54f-7760-4e80-8f46-850a8ce75147-config\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.590309 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92fe1566-b184-4337-b3f6-15f1abd8b0b1-client-ca\") pod \"route-controller-manager-67c6b664cd-bqjgj\" (UID: \"92fe1566-b184-4337-b3f6-15f1abd8b0b1\") " pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.590349 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxvkp\" (UniqueName: \"kubernetes.io/projected/5239d54f-7760-4e80-8f46-850a8ce75147-kube-api-access-bxvkp\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.591770 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5239d54f-7760-4e80-8f46-850a8ce75147-client-ca\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.593670 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92fe1566-b184-4337-b3f6-15f1abd8b0b1-config\") pod \"route-controller-manager-67c6b664cd-bqjgj\" (UID: \"92fe1566-b184-4337-b3f6-15f1abd8b0b1\") " pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.595214 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5239d54f-7760-4e80-8f46-850a8ce75147-proxy-ca-bundles\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.595228 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5239d54f-7760-4e80-8f46-850a8ce75147-config\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.610667 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/92fe1566-b184-4337-b3f6-15f1abd8b0b1-client-ca\") pod \"route-controller-manager-67c6b664cd-bqjgj\" (UID: \"92fe1566-b184-4337-b3f6-15f1abd8b0b1\") " pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.612056 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5239d54f-7760-4e80-8f46-850a8ce75147-serving-cert\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.612574 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/92fe1566-b184-4337-b3f6-15f1abd8b0b1-serving-cert\") pod \"route-controller-manager-67c6b664cd-bqjgj\" (UID: \"92fe1566-b184-4337-b3f6-15f1abd8b0b1\") " pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.616220 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfrmp\" (UniqueName: \"kubernetes.io/projected/92fe1566-b184-4337-b3f6-15f1abd8b0b1-kube-api-access-mfrmp\") pod \"route-controller-manager-67c6b664cd-bqjgj\" (UID: \"92fe1566-b184-4337-b3f6-15f1abd8b0b1\") " pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.619196 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxvkp\" (UniqueName: \"kubernetes.io/projected/5239d54f-7760-4e80-8f46-850a8ce75147-kube-api-access-bxvkp\") pod \"controller-manager-57b6756bff-vmjcv\" (UID: \"5239d54f-7760-4e80-8f46-850a8ce75147\") " pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.642623 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.654914 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:31 crc kubenswrapper[4872]: I0126 09:12:31.891875 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-57b6756bff-vmjcv"] Jan 26 09:12:32 crc kubenswrapper[4872]: I0126 09:12:32.158435 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj"] Jan 26 09:12:32 crc kubenswrapper[4872]: W0126 09:12:32.163593 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod92fe1566_b184_4337_b3f6_15f1abd8b0b1.slice/crio-cc06e96b6e5199f79559410c3179c42e357441dbe46772808dfc3101c269a321 WatchSource:0}: Error finding container cc06e96b6e5199f79559410c3179c42e357441dbe46772808dfc3101c269a321: Status 404 returned error can't find the container with id cc06e96b6e5199f79559410c3179c42e357441dbe46772808dfc3101c269a321 Jan 26 09:12:32 crc kubenswrapper[4872]: I0126 09:12:32.603492 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" event={"ID":"5239d54f-7760-4e80-8f46-850a8ce75147","Type":"ContainerStarted","Data":"5c1cad1f37279a920e44a8fe7642700a26d8c27cece7b0f6099c92035d183730"} Jan 26 09:12:32 crc kubenswrapper[4872]: I0126 09:12:32.603586 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" event={"ID":"5239d54f-7760-4e80-8f46-850a8ce75147","Type":"ContainerStarted","Data":"4057398e8863acb46a35a4a57c79ae759493acfa2c866dec01a8d568d710a7d0"} Jan 26 09:12:32 crc kubenswrapper[4872]: I0126 09:12:32.604283 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:32 crc kubenswrapper[4872]: I0126 09:12:32.605552 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" event={"ID":"92fe1566-b184-4337-b3f6-15f1abd8b0b1","Type":"ContainerStarted","Data":"d1f8256bb802a69180fd5e524191a89e611ea72a246e38a132774898f8816d62"} Jan 26 09:12:32 crc kubenswrapper[4872]: I0126 09:12:32.605588 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" event={"ID":"92fe1566-b184-4337-b3f6-15f1abd8b0b1","Type":"ContainerStarted","Data":"cc06e96b6e5199f79559410c3179c42e357441dbe46772808dfc3101c269a321"} Jan 26 09:12:32 crc kubenswrapper[4872]: I0126 09:12:32.606167 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:32 crc kubenswrapper[4872]: I0126 09:12:32.622486 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" Jan 26 09:12:32 crc kubenswrapper[4872]: I0126 09:12:32.643186 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-57b6756bff-vmjcv" podStartSLOduration=3.643164824 podStartE2EDuration="3.643164824s" podCreationTimestamp="2026-01-26 09:12:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:12:32.640228293 +0000 UTC m=+285.949068104" watchObservedRunningTime="2026-01-26 09:12:32.643164824 +0000 UTC m=+285.952004625" Jan 26 09:12:32 crc kubenswrapper[4872]: I0126 09:12:32.666175 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" podStartSLOduration=3.666152109 podStartE2EDuration="3.666152109s" podCreationTimestamp="2026-01-26 09:12:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:12:32.662854068 +0000 UTC m=+285.971693879" watchObservedRunningTime="2026-01-26 09:12:32.666152109 +0000 UTC m=+285.974991910" Jan 26 09:12:32 crc kubenswrapper[4872]: I0126 09:12:32.789581 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-67c6b664cd-bqjgj" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.271047 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zksvb"] Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.272909 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.309025 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zksvb"] Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.394444 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42827464-079c-4df6-945b-7bd899c16ecb-bound-sa-token\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.394522 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42827464-079c-4df6-945b-7bd899c16ecb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.394574 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.394609 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42827464-079c-4df6-945b-7bd899c16ecb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.394632 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkr6n\" (UniqueName: \"kubernetes.io/projected/42827464-079c-4df6-945b-7bd899c16ecb-kube-api-access-fkr6n\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.394662 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42827464-079c-4df6-945b-7bd899c16ecb-registry-certificates\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.394689 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42827464-079c-4df6-945b-7bd899c16ecb-registry-tls\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.394717 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42827464-079c-4df6-945b-7bd899c16ecb-trusted-ca\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.419302 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.495919 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42827464-079c-4df6-945b-7bd899c16ecb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.495983 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkr6n\" (UniqueName: \"kubernetes.io/projected/42827464-079c-4df6-945b-7bd899c16ecb-kube-api-access-fkr6n\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.496020 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42827464-079c-4df6-945b-7bd899c16ecb-registry-certificates\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.496058 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42827464-079c-4df6-945b-7bd899c16ecb-registry-tls\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.496119 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42827464-079c-4df6-945b-7bd899c16ecb-trusted-ca\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.496146 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42827464-079c-4df6-945b-7bd899c16ecb-bound-sa-token\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.496186 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42827464-079c-4df6-945b-7bd899c16ecb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.496751 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/42827464-079c-4df6-945b-7bd899c16ecb-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.498462 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/42827464-079c-4df6-945b-7bd899c16ecb-registry-certificates\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.499045 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/42827464-079c-4df6-945b-7bd899c16ecb-trusted-ca\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.505373 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/42827464-079c-4df6-945b-7bd899c16ecb-registry-tls\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.505885 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/42827464-079c-4df6-945b-7bd899c16ecb-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.519374 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkr6n\" (UniqueName: \"kubernetes.io/projected/42827464-079c-4df6-945b-7bd899c16ecb-kube-api-access-fkr6n\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.532307 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/42827464-079c-4df6-945b-7bd899c16ecb-bound-sa-token\") pod \"image-registry-66df7c8f76-zksvb\" (UID: \"42827464-079c-4df6-945b-7bd899c16ecb\") " pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:37 crc kubenswrapper[4872]: I0126 09:12:37.605150 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:38 crc kubenswrapper[4872]: I0126 09:12:38.076528 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zksvb"] Jan 26 09:12:38 crc kubenswrapper[4872]: I0126 09:12:38.669362 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" event={"ID":"42827464-079c-4df6-945b-7bd899c16ecb","Type":"ContainerStarted","Data":"420beb6de9e0bb688a396b5f545ffc8de23e728750683d3b4fa04a581557b01a"} Jan 26 09:12:39 crc kubenswrapper[4872]: I0126 09:12:39.681302 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" event={"ID":"42827464-079c-4df6-945b-7bd899c16ecb","Type":"ContainerStarted","Data":"4be195233810e542d57027ba493ceba794dc2ffdc0d08f2fe8d1c0d675d253a7"} Jan 26 09:12:39 crc kubenswrapper[4872]: I0126 09:12:39.682545 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:39 crc kubenswrapper[4872]: I0126 09:12:39.713116 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" podStartSLOduration=2.713092472 podStartE2EDuration="2.713092472s" podCreationTimestamp="2026-01-26 09:12:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:12:39.708335251 +0000 UTC m=+293.017175052" watchObservedRunningTime="2026-01-26 09:12:39.713092472 +0000 UTC m=+293.021932273" Jan 26 09:12:47 crc kubenswrapper[4872]: I0126 09:12:47.011211 4872 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Jan 26 09:12:57 crc kubenswrapper[4872]: I0126 09:12:57.610523 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-zksvb" Jan 26 09:12:57 crc kubenswrapper[4872]: I0126 09:12:57.670296 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p975q"] Jan 26 09:13:22 crc kubenswrapper[4872]: I0126 09:13:22.707078 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" podUID="9aa4a2c3-14ef-42ee-a4cb-261166d7350d" containerName="registry" containerID="cri-o://f82d544f596d31854f94c8c1eaffeeffc99cb646823caa0f58a123c61e06ed93" gracePeriod=30 Jan 26 09:13:22 crc kubenswrapper[4872]: I0126 09:13:22.969071 4872 generic.go:334] "Generic (PLEG): container finished" podID="9aa4a2c3-14ef-42ee-a4cb-261166d7350d" containerID="f82d544f596d31854f94c8c1eaffeeffc99cb646823caa0f58a123c61e06ed93" exitCode=0 Jan 26 09:13:22 crc kubenswrapper[4872]: I0126 09:13:22.969470 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" event={"ID":"9aa4a2c3-14ef-42ee-a4cb-261166d7350d","Type":"ContainerDied","Data":"f82d544f596d31854f94c8c1eaffeeffc99cb646823caa0f58a123c61e06ed93"} Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.158391 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.223016 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-ca-trust-extracted\") pod \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.223096 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-trusted-ca\") pod \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.223221 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-bound-sa-token\") pod \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.223261 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-registry-tls\") pod \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.223620 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.223710 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-installation-pull-secrets\") pod \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.223756 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-registry-certificates\") pod \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.223792 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpjq7\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-kube-api-access-xpjq7\") pod \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\" (UID: \"9aa4a2c3-14ef-42ee-a4cb-261166d7350d\") " Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.224843 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9aa4a2c3-14ef-42ee-a4cb-261166d7350d" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.225341 4872 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-registry-certificates\") on node \"crc\" DevicePath \"\"" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.227347 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9aa4a2c3-14ef-42ee-a4cb-261166d7350d" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.229578 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9aa4a2c3-14ef-42ee-a4cb-261166d7350d" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.229627 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-kube-api-access-xpjq7" (OuterVolumeSpecName: "kube-api-access-xpjq7") pod "9aa4a2c3-14ef-42ee-a4cb-261166d7350d" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d"). InnerVolumeSpecName "kube-api-access-xpjq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.234811 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "9aa4a2c3-14ef-42ee-a4cb-261166d7350d" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.236482 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9aa4a2c3-14ef-42ee-a4cb-261166d7350d" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.240131 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9aa4a2c3-14ef-42ee-a4cb-261166d7350d" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.254213 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9aa4a2c3-14ef-42ee-a4cb-261166d7350d" (UID: "9aa4a2c3-14ef-42ee-a4cb-261166d7350d"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.326649 4872 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-bound-sa-token\") on node \"crc\" DevicePath \"\"" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.326696 4872 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-registry-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.326707 4872 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.326722 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpjq7\" (UniqueName: \"kubernetes.io/projected/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-kube-api-access-xpjq7\") on node \"crc\" DevicePath \"\"" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.326914 4872 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.327539 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9aa4a2c3-14ef-42ee-a4cb-261166d7350d-trusted-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.979295 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" event={"ID":"9aa4a2c3-14ef-42ee-a4cb-261166d7350d","Type":"ContainerDied","Data":"8bcb06d1756f46daf488c8f81628f94a3382dd5a2a51572dce0912d2e353b845"} Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.979399 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-p975q" Jan 26 09:13:23 crc kubenswrapper[4872]: I0126 09:13:23.980033 4872 scope.go:117] "RemoveContainer" containerID="f82d544f596d31854f94c8c1eaffeeffc99cb646823caa0f58a123c61e06ed93" Jan 26 09:13:24 crc kubenswrapper[4872]: I0126 09:13:24.034970 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p975q"] Jan 26 09:13:24 crc kubenswrapper[4872]: I0126 09:13:24.040488 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-p975q"] Jan 26 09:13:25 crc kubenswrapper[4872]: I0126 09:13:25.195962 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aa4a2c3-14ef-42ee-a4cb-261166d7350d" path="/var/lib/kubelet/pods/9aa4a2c3-14ef-42ee-a4cb-261166d7350d/volumes" Jan 26 09:13:40 crc kubenswrapper[4872]: I0126 09:13:40.227500 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:13:40 crc kubenswrapper[4872]: I0126 09:13:40.228314 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:14:10 crc kubenswrapper[4872]: I0126 09:14:10.228375 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:14:10 crc kubenswrapper[4872]: I0126 09:14:10.229317 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:14:40 crc kubenswrapper[4872]: I0126 09:14:40.227373 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:14:40 crc kubenswrapper[4872]: I0126 09:14:40.228022 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:14:40 crc kubenswrapper[4872]: I0126 09:14:40.228076 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:14:40 crc kubenswrapper[4872]: I0126 09:14:40.228732 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fc57e5384dec6c48143f4ceff8720fb743fb9d09433bb2bd3b7ba7d5cc00b27c"} pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 09:14:40 crc kubenswrapper[4872]: I0126 09:14:40.228786 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" containerID="cri-o://fc57e5384dec6c48143f4ceff8720fb743fb9d09433bb2bd3b7ba7d5cc00b27c" gracePeriod=600 Jan 26 09:14:40 crc kubenswrapper[4872]: I0126 09:14:40.503388 4872 generic.go:334] "Generic (PLEG): container finished" podID="fca44d96-a000-4bf2-8283-a937b0192880" containerID="fc57e5384dec6c48143f4ceff8720fb743fb9d09433bb2bd3b7ba7d5cc00b27c" exitCode=0 Jan 26 09:14:40 crc kubenswrapper[4872]: I0126 09:14:40.503478 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerDied","Data":"fc57e5384dec6c48143f4ceff8720fb743fb9d09433bb2bd3b7ba7d5cc00b27c"} Jan 26 09:14:40 crc kubenswrapper[4872]: I0126 09:14:40.503927 4872 scope.go:117] "RemoveContainer" containerID="28dd0adf5ef23faf62c3b054625b35f112e9c3a187505cc6479cc9cd1c86c6bd" Jan 26 09:14:41 crc kubenswrapper[4872]: I0126 09:14:41.513963 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerStarted","Data":"b2ddad1c18cf250920b95ba098a58a140582a8cf9f6b8c8479e40b6449d7e085"} Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.196439 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz"] Jan 26 09:15:00 crc kubenswrapper[4872]: E0126 09:15:00.197490 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa4a2c3-14ef-42ee-a4cb-261166d7350d" containerName="registry" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.197507 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa4a2c3-14ef-42ee-a4cb-261166d7350d" containerName="registry" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.197620 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa4a2c3-14ef-42ee-a4cb-261166d7350d" containerName="registry" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.198146 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.200611 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.200913 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.217625 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz"] Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.353325 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/572664fc-3a7b-4354-b2f3-e69cb5776ec5-secret-volume\") pod \"collect-profiles-29490315-sh2wz\" (UID: \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.353399 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmwcn\" (UniqueName: \"kubernetes.io/projected/572664fc-3a7b-4354-b2f3-e69cb5776ec5-kube-api-access-tmwcn\") pod \"collect-profiles-29490315-sh2wz\" (UID: \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.353462 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/572664fc-3a7b-4354-b2f3-e69cb5776ec5-config-volume\") pod \"collect-profiles-29490315-sh2wz\" (UID: \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.454947 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/572664fc-3a7b-4354-b2f3-e69cb5776ec5-secret-volume\") pod \"collect-profiles-29490315-sh2wz\" (UID: \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.455368 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmwcn\" (UniqueName: \"kubernetes.io/projected/572664fc-3a7b-4354-b2f3-e69cb5776ec5-kube-api-access-tmwcn\") pod \"collect-profiles-29490315-sh2wz\" (UID: \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.455419 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/572664fc-3a7b-4354-b2f3-e69cb5776ec5-config-volume\") pod \"collect-profiles-29490315-sh2wz\" (UID: \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.456483 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/572664fc-3a7b-4354-b2f3-e69cb5776ec5-config-volume\") pod \"collect-profiles-29490315-sh2wz\" (UID: \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.466047 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/572664fc-3a7b-4354-b2f3-e69cb5776ec5-secret-volume\") pod \"collect-profiles-29490315-sh2wz\" (UID: \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.480124 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmwcn\" (UniqueName: \"kubernetes.io/projected/572664fc-3a7b-4354-b2f3-e69cb5776ec5-kube-api-access-tmwcn\") pod \"collect-profiles-29490315-sh2wz\" (UID: \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.522929 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:15:00 crc kubenswrapper[4872]: I0126 09:15:00.780287 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz"] Jan 26 09:15:00 crc kubenswrapper[4872]: W0126 09:15:00.794371 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod572664fc_3a7b_4354_b2f3_e69cb5776ec5.slice/crio-1a1f8a95e66c27a74c45d4c2bab885d476316891b52a7d72a1d176088bbe46f3 WatchSource:0}: Error finding container 1a1f8a95e66c27a74c45d4c2bab885d476316891b52a7d72a1d176088bbe46f3: Status 404 returned error can't find the container with id 1a1f8a95e66c27a74c45d4c2bab885d476316891b52a7d72a1d176088bbe46f3 Jan 26 09:15:01 crc kubenswrapper[4872]: I0126 09:15:01.641052 4872 generic.go:334] "Generic (PLEG): container finished" podID="572664fc-3a7b-4354-b2f3-e69cb5776ec5" containerID="2097e0c89b8371da27960a8e6fba56fbc24bf7a830fad510258e712ffe07a9ff" exitCode=0 Jan 26 09:15:01 crc kubenswrapper[4872]: I0126 09:15:01.641110 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" event={"ID":"572664fc-3a7b-4354-b2f3-e69cb5776ec5","Type":"ContainerDied","Data":"2097e0c89b8371da27960a8e6fba56fbc24bf7a830fad510258e712ffe07a9ff"} Jan 26 09:15:01 crc kubenswrapper[4872]: I0126 09:15:01.641150 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" event={"ID":"572664fc-3a7b-4354-b2f3-e69cb5776ec5","Type":"ContainerStarted","Data":"1a1f8a95e66c27a74c45d4c2bab885d476316891b52a7d72a1d176088bbe46f3"} Jan 26 09:15:02 crc kubenswrapper[4872]: I0126 09:15:02.855892 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:15:02 crc kubenswrapper[4872]: I0126 09:15:02.990707 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/572664fc-3a7b-4354-b2f3-e69cb5776ec5-secret-volume\") pod \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\" (UID: \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\") " Jan 26 09:15:02 crc kubenswrapper[4872]: I0126 09:15:02.991064 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmwcn\" (UniqueName: \"kubernetes.io/projected/572664fc-3a7b-4354-b2f3-e69cb5776ec5-kube-api-access-tmwcn\") pod \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\" (UID: \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\") " Jan 26 09:15:02 crc kubenswrapper[4872]: I0126 09:15:02.991168 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/572664fc-3a7b-4354-b2f3-e69cb5776ec5-config-volume\") pod \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\" (UID: \"572664fc-3a7b-4354-b2f3-e69cb5776ec5\") " Jan 26 09:15:02 crc kubenswrapper[4872]: I0126 09:15:02.992512 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/572664fc-3a7b-4354-b2f3-e69cb5776ec5-config-volume" (OuterVolumeSpecName: "config-volume") pod "572664fc-3a7b-4354-b2f3-e69cb5776ec5" (UID: "572664fc-3a7b-4354-b2f3-e69cb5776ec5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:15:02 crc kubenswrapper[4872]: I0126 09:15:02.998281 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/572664fc-3a7b-4354-b2f3-e69cb5776ec5-kube-api-access-tmwcn" (OuterVolumeSpecName: "kube-api-access-tmwcn") pod "572664fc-3a7b-4354-b2f3-e69cb5776ec5" (UID: "572664fc-3a7b-4354-b2f3-e69cb5776ec5"). InnerVolumeSpecName "kube-api-access-tmwcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:15:02 crc kubenswrapper[4872]: I0126 09:15:02.998311 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/572664fc-3a7b-4354-b2f3-e69cb5776ec5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "572664fc-3a7b-4354-b2f3-e69cb5776ec5" (UID: "572664fc-3a7b-4354-b2f3-e69cb5776ec5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:15:03 crc kubenswrapper[4872]: I0126 09:15:03.093327 4872 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/572664fc-3a7b-4354-b2f3-e69cb5776ec5-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 09:15:03 crc kubenswrapper[4872]: I0126 09:15:03.093390 4872 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/572664fc-3a7b-4354-b2f3-e69cb5776ec5-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 09:15:03 crc kubenswrapper[4872]: I0126 09:15:03.093405 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmwcn\" (UniqueName: \"kubernetes.io/projected/572664fc-3a7b-4354-b2f3-e69cb5776ec5-kube-api-access-tmwcn\") on node \"crc\" DevicePath \"\"" Jan 26 09:15:03 crc kubenswrapper[4872]: I0126 09:15:03.657488 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" event={"ID":"572664fc-3a7b-4354-b2f3-e69cb5776ec5","Type":"ContainerDied","Data":"1a1f8a95e66c27a74c45d4c2bab885d476316891b52a7d72a1d176088bbe46f3"} Jan 26 09:15:03 crc kubenswrapper[4872]: I0126 09:15:03.657567 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a1f8a95e66c27a74c45d4c2bab885d476316891b52a7d72a1d176088bbe46f3" Jan 26 09:15:03 crc kubenswrapper[4872]: I0126 09:15:03.657669 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490315-sh2wz" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.489880 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-jnspw"] Jan 26 09:16:20 crc kubenswrapper[4872]: E0126 09:16:20.491829 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572664fc-3a7b-4354-b2f3-e69cb5776ec5" containerName="collect-profiles" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.491917 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="572664fc-3a7b-4354-b2f3-e69cb5776ec5" containerName="collect-profiles" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.492096 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="572664fc-3a7b-4354-b2f3-e69cb5776ec5" containerName="collect-profiles" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.492679 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-jnspw" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.496162 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.496580 4872 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-fkc2h" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.496625 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.500206 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-jnspw"] Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.504714 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-m97zb"] Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.505793 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-m97zb" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.508264 4872 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-wzlbd" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.513330 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-hxz6l"] Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.514247 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-hxz6l" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.518618 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-m97zb"] Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.521423 4872 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-6s9cr" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.538646 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-hxz6l"] Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.578384 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78hvb\" (UniqueName: \"kubernetes.io/projected/e165986a-e1a8-47bc-8e3c-2609509f1ede-kube-api-access-78hvb\") pod \"cert-manager-webhook-687f57d79b-hxz6l\" (UID: \"e165986a-e1a8-47bc-8e3c-2609509f1ede\") " pod="cert-manager/cert-manager-webhook-687f57d79b-hxz6l" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.578481 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn9g4\" (UniqueName: \"kubernetes.io/projected/1d91c9be-4442-4976-8714-bf07b8777576-kube-api-access-kn9g4\") pod \"cert-manager-858654f9db-m97zb\" (UID: \"1d91c9be-4442-4976-8714-bf07b8777576\") " pod="cert-manager/cert-manager-858654f9db-m97zb" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.578624 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcpn6\" (UniqueName: \"kubernetes.io/projected/f1d545e1-0b8c-481e-a979-760d92eec6da-kube-api-access-qcpn6\") pod \"cert-manager-cainjector-cf98fcc89-jnspw\" (UID: \"f1d545e1-0b8c-481e-a979-760d92eec6da\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-jnspw" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.680244 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcpn6\" (UniqueName: \"kubernetes.io/projected/f1d545e1-0b8c-481e-a979-760d92eec6da-kube-api-access-qcpn6\") pod \"cert-manager-cainjector-cf98fcc89-jnspw\" (UID: \"f1d545e1-0b8c-481e-a979-760d92eec6da\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-jnspw" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.680305 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78hvb\" (UniqueName: \"kubernetes.io/projected/e165986a-e1a8-47bc-8e3c-2609509f1ede-kube-api-access-78hvb\") pod \"cert-manager-webhook-687f57d79b-hxz6l\" (UID: \"e165986a-e1a8-47bc-8e3c-2609509f1ede\") " pod="cert-manager/cert-manager-webhook-687f57d79b-hxz6l" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.680338 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn9g4\" (UniqueName: \"kubernetes.io/projected/1d91c9be-4442-4976-8714-bf07b8777576-kube-api-access-kn9g4\") pod \"cert-manager-858654f9db-m97zb\" (UID: \"1d91c9be-4442-4976-8714-bf07b8777576\") " pod="cert-manager/cert-manager-858654f9db-m97zb" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.706415 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn9g4\" (UniqueName: \"kubernetes.io/projected/1d91c9be-4442-4976-8714-bf07b8777576-kube-api-access-kn9g4\") pod \"cert-manager-858654f9db-m97zb\" (UID: \"1d91c9be-4442-4976-8714-bf07b8777576\") " pod="cert-manager/cert-manager-858654f9db-m97zb" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.706416 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78hvb\" (UniqueName: \"kubernetes.io/projected/e165986a-e1a8-47bc-8e3c-2609509f1ede-kube-api-access-78hvb\") pod \"cert-manager-webhook-687f57d79b-hxz6l\" (UID: \"e165986a-e1a8-47bc-8e3c-2609509f1ede\") " pod="cert-manager/cert-manager-webhook-687f57d79b-hxz6l" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.709287 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcpn6\" (UniqueName: \"kubernetes.io/projected/f1d545e1-0b8c-481e-a979-760d92eec6da-kube-api-access-qcpn6\") pod \"cert-manager-cainjector-cf98fcc89-jnspw\" (UID: \"f1d545e1-0b8c-481e-a979-760d92eec6da\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-jnspw" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.816716 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-jnspw" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.824727 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-m97zb" Jan 26 09:16:20 crc kubenswrapper[4872]: I0126 09:16:20.832360 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-hxz6l" Jan 26 09:16:21 crc kubenswrapper[4872]: I0126 09:16:21.097497 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-hxz6l"] Jan 26 09:16:21 crc kubenswrapper[4872]: I0126 09:16:21.127324 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 09:16:21 crc kubenswrapper[4872]: I0126 09:16:21.240298 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-hxz6l" event={"ID":"e165986a-e1a8-47bc-8e3c-2609509f1ede","Type":"ContainerStarted","Data":"485d9383c4b8450f4254a4498e00eea157d7c5f55d5409d299aaa45a61867174"} Jan 26 09:16:21 crc kubenswrapper[4872]: I0126 09:16:21.352329 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-m97zb"] Jan 26 09:16:21 crc kubenswrapper[4872]: W0126 09:16:21.358775 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d91c9be_4442_4976_8714_bf07b8777576.slice/crio-7c01eb78594358154460db7203b931e9d8dfa1883e38af0f1047f5ce8434a9a6 WatchSource:0}: Error finding container 7c01eb78594358154460db7203b931e9d8dfa1883e38af0f1047f5ce8434a9a6: Status 404 returned error can't find the container with id 7c01eb78594358154460db7203b931e9d8dfa1883e38af0f1047f5ce8434a9a6 Jan 26 09:16:21 crc kubenswrapper[4872]: I0126 09:16:21.373415 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-jnspw"] Jan 26 09:16:21 crc kubenswrapper[4872]: W0126 09:16:21.378482 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1d545e1_0b8c_481e_a979_760d92eec6da.slice/crio-840ea867170b9c319027fbe9c3bbbc36784014312eafa446a1478716b3b48b6c WatchSource:0}: Error finding container 840ea867170b9c319027fbe9c3bbbc36784014312eafa446a1478716b3b48b6c: Status 404 returned error can't find the container with id 840ea867170b9c319027fbe9c3bbbc36784014312eafa446a1478716b3b48b6c Jan 26 09:16:22 crc kubenswrapper[4872]: I0126 09:16:22.253422 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-m97zb" event={"ID":"1d91c9be-4442-4976-8714-bf07b8777576","Type":"ContainerStarted","Data":"7c01eb78594358154460db7203b931e9d8dfa1883e38af0f1047f5ce8434a9a6"} Jan 26 09:16:22 crc kubenswrapper[4872]: I0126 09:16:22.255301 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-jnspw" event={"ID":"f1d545e1-0b8c-481e-a979-760d92eec6da","Type":"ContainerStarted","Data":"840ea867170b9c319027fbe9c3bbbc36784014312eafa446a1478716b3b48b6c"} Jan 26 09:16:24 crc kubenswrapper[4872]: I0126 09:16:24.267886 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-hxz6l" event={"ID":"e165986a-e1a8-47bc-8e3c-2609509f1ede","Type":"ContainerStarted","Data":"1538e6d19c1df966e658e13a2e7aab01dff7188ae2f307f978da6c9a3cc6f47e"} Jan 26 09:16:24 crc kubenswrapper[4872]: I0126 09:16:24.268304 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-hxz6l" Jan 26 09:16:24 crc kubenswrapper[4872]: I0126 09:16:24.291766 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-hxz6l" podStartSLOduration=1.9963145450000002 podStartE2EDuration="4.291743655s" podCreationTimestamp="2026-01-26 09:16:20 +0000 UTC" firstStartedPulling="2026-01-26 09:16:21.127105853 +0000 UTC m=+514.435945654" lastFinishedPulling="2026-01-26 09:16:23.422534963 +0000 UTC m=+516.731374764" observedRunningTime="2026-01-26 09:16:24.28997993 +0000 UTC m=+517.598819731" watchObservedRunningTime="2026-01-26 09:16:24.291743655 +0000 UTC m=+517.600583456" Jan 26 09:16:25 crc kubenswrapper[4872]: I0126 09:16:25.277540 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-jnspw" event={"ID":"f1d545e1-0b8c-481e-a979-760d92eec6da","Type":"ContainerStarted","Data":"a8ddd79a24c041619299d8fdcd6da868626e808baa2e4dbca59803ee2e124605"} Jan 26 09:16:26 crc kubenswrapper[4872]: I0126 09:16:26.285557 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-m97zb" event={"ID":"1d91c9be-4442-4976-8714-bf07b8777576","Type":"ContainerStarted","Data":"5ca1dd6ffb35c3833ab66ed5fddfd4a7653adf717923d7e8ade4ca31a81c058b"} Jan 26 09:16:26 crc kubenswrapper[4872]: I0126 09:16:26.322366 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-m97zb" podStartSLOduration=2.358443761 podStartE2EDuration="6.322336121s" podCreationTimestamp="2026-01-26 09:16:20 +0000 UTC" firstStartedPulling="2026-01-26 09:16:21.361750669 +0000 UTC m=+514.670590470" lastFinishedPulling="2026-01-26 09:16:25.325643029 +0000 UTC m=+518.634482830" observedRunningTime="2026-01-26 09:16:26.306265856 +0000 UTC m=+519.615105657" watchObservedRunningTime="2026-01-26 09:16:26.322336121 +0000 UTC m=+519.631175932" Jan 26 09:16:26 crc kubenswrapper[4872]: I0126 09:16:26.323205 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-jnspw" podStartSLOduration=2.630433551 podStartE2EDuration="6.323198283s" podCreationTimestamp="2026-01-26 09:16:20 +0000 UTC" firstStartedPulling="2026-01-26 09:16:21.381641312 +0000 UTC m=+514.690481113" lastFinishedPulling="2026-01-26 09:16:25.074406004 +0000 UTC m=+518.383245845" observedRunningTime="2026-01-26 09:16:26.320974016 +0000 UTC m=+519.629813837" watchObservedRunningTime="2026-01-26 09:16:26.323198283 +0000 UTC m=+519.632038084" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.131047 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6p5bj"] Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.131812 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovn-controller" containerID="cri-o://f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad" gracePeriod=30 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.131860 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="sbdb" containerID="cri-o://1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665" gracePeriod=30 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.131943 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="nbdb" containerID="cri-o://72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6" gracePeriod=30 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.131975 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="northd" containerID="cri-o://6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7" gracePeriod=30 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.132030 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovn-acl-logging" containerID="cri-o://b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9" gracePeriod=30 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.132015 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="kube-rbac-proxy-node" containerID="cri-o://41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc" gracePeriod=30 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.131911 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65" gracePeriod=30 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.168929 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" containerID="cri-o://41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed" gracePeriod=30 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.313146 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovnkube-controller/2.log" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.315739 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovn-acl-logging/0.log" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.316400 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovn-controller/0.log" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.316828 4872 generic.go:334] "Generic (PLEG): container finished" podID="e2577331-1886-4abb-896b-8d8a112e40d8" containerID="41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed" exitCode=0 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.316887 4872 generic.go:334] "Generic (PLEG): container finished" podID="e2577331-1886-4abb-896b-8d8a112e40d8" containerID="b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65" exitCode=0 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.316896 4872 generic.go:334] "Generic (PLEG): container finished" podID="e2577331-1886-4abb-896b-8d8a112e40d8" containerID="41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc" exitCode=0 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.316903 4872 generic.go:334] "Generic (PLEG): container finished" podID="e2577331-1886-4abb-896b-8d8a112e40d8" containerID="b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9" exitCode=143 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.316910 4872 generic.go:334] "Generic (PLEG): container finished" podID="e2577331-1886-4abb-896b-8d8a112e40d8" containerID="f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad" exitCode=143 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.316949 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed"} Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.316982 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65"} Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.316992 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc"} Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.317001 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9"} Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.317010 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad"} Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.317027 4872 scope.go:117] "RemoveContainer" containerID="e3256926557c4db912bdd31fcdfb473e48c0de635b1367ade5d07c4e9429e621" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.320298 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-swcf5_3579107d-a172-4905-a491-6dd2b5a304a5/kube-multus/1.log" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.320759 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-swcf5_3579107d-a172-4905-a491-6dd2b5a304a5/kube-multus/0.log" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.320791 4872 generic.go:334] "Generic (PLEG): container finished" podID="3579107d-a172-4905-a491-6dd2b5a304a5" containerID="4960fa5dfa153921fd4c32dffe9662db8304b2a8242ef2b570ccc174c76279d8" exitCode=2 Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.320847 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-swcf5" event={"ID":"3579107d-a172-4905-a491-6dd2b5a304a5","Type":"ContainerDied","Data":"4960fa5dfa153921fd4c32dffe9662db8304b2a8242ef2b570ccc174c76279d8"} Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.321430 4872 scope.go:117] "RemoveContainer" containerID="4960fa5dfa153921fd4c32dffe9662db8304b2a8242ef2b570ccc174c76279d8" Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.321619 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-swcf5_openshift-multus(3579107d-a172-4905-a491-6dd2b5a304a5)\"" pod="openshift-multus/multus-swcf5" podUID="3579107d-a172-4905-a491-6dd2b5a304a5" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.427614 4872 scope.go:117] "RemoveContainer" containerID="27a07793c84ece24c6a36c8e7e96b22f6867350159af781d924fd1f7842da924" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.475527 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovn-acl-logging/0.log" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.476122 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovn-controller/0.log" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.476552 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536035 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cqtgt"] Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.536337 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="sbdb" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536352 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="sbdb" Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.536361 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="kubecfg-setup" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536367 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="kubecfg-setup" Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.536376 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="nbdb" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536386 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="nbdb" Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.536400 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovn-acl-logging" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536409 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovn-acl-logging" Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.536418 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536424 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.536431 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="kube-rbac-proxy-node" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536437 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="kube-rbac-proxy-node" Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.536447 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovn-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536453 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovn-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.536466 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="kube-rbac-proxy-ovn-metrics" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536474 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="kube-rbac-proxy-ovn-metrics" Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.536486 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="northd" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536493 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="northd" Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.536502 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536508 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.536515 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536521 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536652 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536666 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="kube-rbac-proxy-ovn-metrics" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536674 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="northd" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536685 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536693 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovn-acl-logging" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536701 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="kube-rbac-proxy-node" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536710 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536720 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="nbdb" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536727 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="sbdb" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536734 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovn-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: E0126 09:16:30.536857 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536865 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.536961 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" containerName="ovnkube-controller" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.538746 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548079 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-etc-openvswitch\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548135 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-log-socket\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548169 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e2577331-1886-4abb-896b-8d8a112e40d8-ovn-node-metrics-cert\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548212 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtxcp\" (UniqueName: \"kubernetes.io/projected/e2577331-1886-4abb-896b-8d8a112e40d8-kube-api-access-qtxcp\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548246 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-systemd-units\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548268 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-slash\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548292 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-ovnkube-config\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548312 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-cni-netd\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548329 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548350 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-kubelet\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548387 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-cni-bin\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548406 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-openvswitch\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548475 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-node-log\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548502 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-ovnkube-script-lib\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548546 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-ovn\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548601 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-var-lib-openvswitch\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548629 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-run-ovn-kubernetes\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548655 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-systemd\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548696 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-env-overrides\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.548717 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-run-netns\") pod \"e2577331-1886-4abb-896b-8d8a112e40d8\" (UID: \"e2577331-1886-4abb-896b-8d8a112e40d8\") " Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.549011 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.549047 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.549067 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-log-socket" (OuterVolumeSpecName: "log-socket") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552034 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552111 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552135 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-slash" (OuterVolumeSpecName: "host-slash") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552361 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552421 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552419 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552452 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552385 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-node-log" (OuterVolumeSpecName: "node-log") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552493 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552493 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552521 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552848 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552897 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.552945 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.558666 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2577331-1886-4abb-896b-8d8a112e40d8-kube-api-access-qtxcp" (OuterVolumeSpecName: "kube-api-access-qtxcp") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "kube-api-access-qtxcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.558849 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2577331-1886-4abb-896b-8d8a112e40d8-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.575423 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "e2577331-1886-4abb-896b-8d8a112e40d8" (UID: "e2577331-1886-4abb-896b-8d8a112e40d8"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.650434 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.650492 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-run-netns\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.650515 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68e8ae60-4d44-412d-be0a-575f475d5117-env-overrides\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.650536 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnsb9\" (UniqueName: \"kubernetes.io/projected/68e8ae60-4d44-412d-be0a-575f475d5117-kube-api-access-nnsb9\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.650731 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-var-lib-openvswitch\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.650916 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/68e8ae60-4d44-412d-be0a-575f475d5117-ovnkube-script-lib\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.650987 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-kubelet\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651036 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-slash\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651158 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68e8ae60-4d44-412d-be0a-575f475d5117-ovnkube-config\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651208 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-log-socket\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651273 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-run-systemd\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651319 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-node-log\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651372 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-run-ovn-kubernetes\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651432 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-systemd-units\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651467 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-run-ovn\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651498 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-cni-bin\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651545 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-etc-openvswitch\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651576 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68e8ae60-4d44-412d-be0a-575f475d5117-ovn-node-metrics-cert\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651632 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-cni-netd\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651727 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-run-openvswitch\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651840 4872 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-node-log\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651858 4872 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651878 4872 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651890 4872 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651902 4872 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651911 4872 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-systemd\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651921 4872 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-env-overrides\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651930 4872 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-run-netns\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651942 4872 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651952 4872 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-log-socket\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651963 4872 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e2577331-1886-4abb-896b-8d8a112e40d8-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651974 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtxcp\" (UniqueName: \"kubernetes.io/projected/e2577331-1886-4abb-896b-8d8a112e40d8-kube-api-access-qtxcp\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651987 4872 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-systemd-units\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.651998 4872 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-slash\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.652010 4872 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e2577331-1886-4abb-896b-8d8a112e40d8-ovnkube-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.652021 4872 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-cni-netd\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.652031 4872 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.652041 4872 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-kubelet\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.652052 4872 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-host-cni-bin\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.652060 4872 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e2577331-1886-4abb-896b-8d8a112e40d8-run-openvswitch\") on node \"crc\" DevicePath \"\"" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753351 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-run-netns\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753415 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68e8ae60-4d44-412d-be0a-575f475d5117-env-overrides\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753443 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnsb9\" (UniqueName: \"kubernetes.io/projected/68e8ae60-4d44-412d-be0a-575f475d5117-kube-api-access-nnsb9\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753469 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-var-lib-openvswitch\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753491 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/68e8ae60-4d44-412d-be0a-575f475d5117-ovnkube-script-lib\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753521 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-kubelet\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753539 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-slash\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753581 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68e8ae60-4d44-412d-be0a-575f475d5117-ovnkube-config\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753581 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-run-netns\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753607 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-log-socket\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753664 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-log-socket\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753668 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-kubelet\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753692 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-var-lib-openvswitch\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.753733 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-slash\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.754479 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68e8ae60-4d44-412d-be0a-575f475d5117-env-overrides\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.754619 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-run-systemd\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.754684 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68e8ae60-4d44-412d-be0a-575f475d5117-ovnkube-config\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.754690 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-run-systemd\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.754845 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-node-log\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.754869 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-run-ovn-kubernetes\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.754896 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-systemd-units\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.754916 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-run-ovn\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.754933 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-cni-bin\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.754952 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-etc-openvswitch\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.754971 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68e8ae60-4d44-412d-be0a-575f475d5117-ovn-node-metrics-cert\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.754990 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-cni-netd\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.755009 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-run-openvswitch\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.755030 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.755028 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-node-log\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.755085 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.755120 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-run-ovn-kubernetes\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.755148 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-run-openvswitch\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.755150 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-cni-netd\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.755156 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-etc-openvswitch\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.755191 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-systemd-units\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.755171 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-run-ovn\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.755221 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/68e8ae60-4d44-412d-be0a-575f475d5117-host-cni-bin\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.755416 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/68e8ae60-4d44-412d-be0a-575f475d5117-ovnkube-script-lib\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.763527 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68e8ae60-4d44-412d-be0a-575f475d5117-ovn-node-metrics-cert\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.780299 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnsb9\" (UniqueName: \"kubernetes.io/projected/68e8ae60-4d44-412d-be0a-575f475d5117-kube-api-access-nnsb9\") pod \"ovnkube-node-cqtgt\" (UID: \"68e8ae60-4d44-412d-be0a-575f475d5117\") " pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.835863 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-hxz6l" Jan 26 09:16:30 crc kubenswrapper[4872]: I0126 09:16:30.858244 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.331717 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-swcf5_3579107d-a172-4905-a491-6dd2b5a304a5/kube-multus/1.log" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.338650 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovn-acl-logging/0.log" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.339328 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-6p5bj_e2577331-1886-4abb-896b-8d8a112e40d8/ovn-controller/0.log" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.339731 4872 generic.go:334] "Generic (PLEG): container finished" podID="e2577331-1886-4abb-896b-8d8a112e40d8" containerID="1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665" exitCode=0 Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.339770 4872 generic.go:334] "Generic (PLEG): container finished" podID="e2577331-1886-4abb-896b-8d8a112e40d8" containerID="72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6" exitCode=0 Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.339782 4872 generic.go:334] "Generic (PLEG): container finished" podID="e2577331-1886-4abb-896b-8d8a112e40d8" containerID="6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7" exitCode=0 Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.339885 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.339867 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665"} Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.340059 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6"} Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.340108 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7"} Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.340126 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-6p5bj" event={"ID":"e2577331-1886-4abb-896b-8d8a112e40d8","Type":"ContainerDied","Data":"457a75a10afa7862edbcafd63b2fcc0a7ba8b01cb2bd81e8340eea6fba0e71af"} Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.340133 4872 scope.go:117] "RemoveContainer" containerID="41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.353376 4872 generic.go:334] "Generic (PLEG): container finished" podID="68e8ae60-4d44-412d-be0a-575f475d5117" containerID="b3bf83b1f8f21474980e4887bdd7c4ad62bddc35c2250fa3e34db723f563e449" exitCode=0 Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.353455 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" event={"ID":"68e8ae60-4d44-412d-be0a-575f475d5117","Type":"ContainerDied","Data":"b3bf83b1f8f21474980e4887bdd7c4ad62bddc35c2250fa3e34db723f563e449"} Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.353492 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" event={"ID":"68e8ae60-4d44-412d-be0a-575f475d5117","Type":"ContainerStarted","Data":"d79ed8ce0f3e69ba8b3c379053e4834aced9317acac2d711338b88513e1511d6"} Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.375056 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6p5bj"] Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.380828 4872 scope.go:117] "RemoveContainer" containerID="1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.383029 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-6p5bj"] Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.401133 4872 scope.go:117] "RemoveContainer" containerID="72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.437925 4872 scope.go:117] "RemoveContainer" containerID="6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.458298 4872 scope.go:117] "RemoveContainer" containerID="b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.491849 4872 scope.go:117] "RemoveContainer" containerID="41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.516008 4872 scope.go:117] "RemoveContainer" containerID="b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.535748 4872 scope.go:117] "RemoveContainer" containerID="f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.557674 4872 scope.go:117] "RemoveContainer" containerID="5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.577301 4872 scope.go:117] "RemoveContainer" containerID="41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed" Jan 26 09:16:31 crc kubenswrapper[4872]: E0126 09:16:31.578068 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed\": container with ID starting with 41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed not found: ID does not exist" containerID="41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.578106 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed"} err="failed to get container status \"41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed\": rpc error: code = NotFound desc = could not find container \"41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed\": container with ID starting with 41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.578151 4872 scope.go:117] "RemoveContainer" containerID="1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665" Jan 26 09:16:31 crc kubenswrapper[4872]: E0126 09:16:31.578582 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\": container with ID starting with 1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665 not found: ID does not exist" containerID="1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.578605 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665"} err="failed to get container status \"1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\": rpc error: code = NotFound desc = could not find container \"1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\": container with ID starting with 1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.578621 4872 scope.go:117] "RemoveContainer" containerID="72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6" Jan 26 09:16:31 crc kubenswrapper[4872]: E0126 09:16:31.578933 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\": container with ID starting with 72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6 not found: ID does not exist" containerID="72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.578952 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6"} err="failed to get container status \"72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\": rpc error: code = NotFound desc = could not find container \"72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\": container with ID starting with 72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.578965 4872 scope.go:117] "RemoveContainer" containerID="6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7" Jan 26 09:16:31 crc kubenswrapper[4872]: E0126 09:16:31.579166 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\": container with ID starting with 6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7 not found: ID does not exist" containerID="6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.579190 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7"} err="failed to get container status \"6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\": rpc error: code = NotFound desc = could not find container \"6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\": container with ID starting with 6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.579204 4872 scope.go:117] "RemoveContainer" containerID="b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65" Jan 26 09:16:31 crc kubenswrapper[4872]: E0126 09:16:31.579610 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\": container with ID starting with b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65 not found: ID does not exist" containerID="b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.579628 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65"} err="failed to get container status \"b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\": rpc error: code = NotFound desc = could not find container \"b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\": container with ID starting with b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.579641 4872 scope.go:117] "RemoveContainer" containerID="41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc" Jan 26 09:16:31 crc kubenswrapper[4872]: E0126 09:16:31.579844 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\": container with ID starting with 41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc not found: ID does not exist" containerID="41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.579868 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc"} err="failed to get container status \"41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\": rpc error: code = NotFound desc = could not find container \"41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\": container with ID starting with 41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.579881 4872 scope.go:117] "RemoveContainer" containerID="b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9" Jan 26 09:16:31 crc kubenswrapper[4872]: E0126 09:16:31.580202 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\": container with ID starting with b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9 not found: ID does not exist" containerID="b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.580225 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9"} err="failed to get container status \"b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\": rpc error: code = NotFound desc = could not find container \"b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\": container with ID starting with b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.580237 4872 scope.go:117] "RemoveContainer" containerID="f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad" Jan 26 09:16:31 crc kubenswrapper[4872]: E0126 09:16:31.580956 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\": container with ID starting with f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad not found: ID does not exist" containerID="f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.580974 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad"} err="failed to get container status \"f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\": rpc error: code = NotFound desc = could not find container \"f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\": container with ID starting with f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.580988 4872 scope.go:117] "RemoveContainer" containerID="5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c" Jan 26 09:16:31 crc kubenswrapper[4872]: E0126 09:16:31.581325 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\": container with ID starting with 5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c not found: ID does not exist" containerID="5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.581346 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c"} err="failed to get container status \"5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\": rpc error: code = NotFound desc = could not find container \"5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\": container with ID starting with 5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.581364 4872 scope.go:117] "RemoveContainer" containerID="41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.581629 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed"} err="failed to get container status \"41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed\": rpc error: code = NotFound desc = could not find container \"41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed\": container with ID starting with 41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.581676 4872 scope.go:117] "RemoveContainer" containerID="1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.581984 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665"} err="failed to get container status \"1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\": rpc error: code = NotFound desc = could not find container \"1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\": container with ID starting with 1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.582004 4872 scope.go:117] "RemoveContainer" containerID="72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.582325 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6"} err="failed to get container status \"72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\": rpc error: code = NotFound desc = could not find container \"72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\": container with ID starting with 72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.582346 4872 scope.go:117] "RemoveContainer" containerID="6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.582689 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7"} err="failed to get container status \"6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\": rpc error: code = NotFound desc = could not find container \"6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\": container with ID starting with 6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.582717 4872 scope.go:117] "RemoveContainer" containerID="b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.583118 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65"} err="failed to get container status \"b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\": rpc error: code = NotFound desc = could not find container \"b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\": container with ID starting with b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.583187 4872 scope.go:117] "RemoveContainer" containerID="41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.583609 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc"} err="failed to get container status \"41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\": rpc error: code = NotFound desc = could not find container \"41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\": container with ID starting with 41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.583630 4872 scope.go:117] "RemoveContainer" containerID="b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.583961 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9"} err="failed to get container status \"b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\": rpc error: code = NotFound desc = could not find container \"b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\": container with ID starting with b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.583987 4872 scope.go:117] "RemoveContainer" containerID="f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.584321 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad"} err="failed to get container status \"f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\": rpc error: code = NotFound desc = could not find container \"f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\": container with ID starting with f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.584341 4872 scope.go:117] "RemoveContainer" containerID="5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.584703 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c"} err="failed to get container status \"5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\": rpc error: code = NotFound desc = could not find container \"5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\": container with ID starting with 5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.584728 4872 scope.go:117] "RemoveContainer" containerID="41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.585061 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed"} err="failed to get container status \"41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed\": rpc error: code = NotFound desc = could not find container \"41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed\": container with ID starting with 41654794b8ba61c4ee77c14bdd75260838cd634e1dd88fb33c12fdb06f415bed not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.585089 4872 scope.go:117] "RemoveContainer" containerID="1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.585402 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665"} err="failed to get container status \"1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\": rpc error: code = NotFound desc = could not find container \"1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665\": container with ID starting with 1ad72bc4e76e2a144a9fe346398d3a48c8e479b137b01a8686dbaa50d5060665 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.585438 4872 scope.go:117] "RemoveContainer" containerID="72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.585821 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6"} err="failed to get container status \"72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\": rpc error: code = NotFound desc = could not find container \"72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6\": container with ID starting with 72d258bfa528cbb4623dc667d3915836480b84cb7617377d70ade640776063c6 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.585868 4872 scope.go:117] "RemoveContainer" containerID="6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.586285 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7"} err="failed to get container status \"6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\": rpc error: code = NotFound desc = could not find container \"6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7\": container with ID starting with 6fd479fb536284a784ac0619ac7a8e9aa771b05b5593bf02103402ea6a9f76b7 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.586323 4872 scope.go:117] "RemoveContainer" containerID="b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.586630 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65"} err="failed to get container status \"b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\": rpc error: code = NotFound desc = could not find container \"b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65\": container with ID starting with b69a12623218158138f0b34f3e20ffaeab4565b89748a1b786614c521534aa65 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.586650 4872 scope.go:117] "RemoveContainer" containerID="41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.587023 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc"} err="failed to get container status \"41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\": rpc error: code = NotFound desc = could not find container \"41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc\": container with ID starting with 41e43cf58859bcb7b62fbe8fbbe8788deb72a29575215b0f71d0c998891806bc not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.587050 4872 scope.go:117] "RemoveContainer" containerID="b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.587390 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9"} err="failed to get container status \"b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\": rpc error: code = NotFound desc = could not find container \"b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9\": container with ID starting with b613b355654e9a8d53ecca720b70db2ffa71ab0647e6434c868cf7b431e1c3a9 not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.587423 4872 scope.go:117] "RemoveContainer" containerID="f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.587982 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad"} err="failed to get container status \"f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\": rpc error: code = NotFound desc = could not find container \"f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad\": container with ID starting with f2327de50840750f8c8accf8c1748ee08aaa42d286f12ad768731196d8bc24ad not found: ID does not exist" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.588003 4872 scope.go:117] "RemoveContainer" containerID="5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c" Jan 26 09:16:31 crc kubenswrapper[4872]: I0126 09:16:31.588333 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c"} err="failed to get container status \"5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\": rpc error: code = NotFound desc = could not find container \"5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c\": container with ID starting with 5744c3ed0fd273e0b1c19f5daca01b6f3358b9c0841233e7dc82d925734e0a1c not found: ID does not exist" Jan 26 09:16:32 crc kubenswrapper[4872]: I0126 09:16:32.363265 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" event={"ID":"68e8ae60-4d44-412d-be0a-575f475d5117","Type":"ContainerStarted","Data":"c1923c0b6797dbe95acade5eaa3a17a1b6a5fd2b4fd7c96580236c742a0f6969"} Jan 26 09:16:32 crc kubenswrapper[4872]: I0126 09:16:32.363735 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" event={"ID":"68e8ae60-4d44-412d-be0a-575f475d5117","Type":"ContainerStarted","Data":"7d5cb25692a7684a6fa2a052e75bf695fb7a3425e69aecc2ecb842635d5580f1"} Jan 26 09:16:32 crc kubenswrapper[4872]: I0126 09:16:32.363753 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" event={"ID":"68e8ae60-4d44-412d-be0a-575f475d5117","Type":"ContainerStarted","Data":"711507b4c5f2d4c87869612e525f3f99bf86294013d0267946de5c3071b29a9e"} Jan 26 09:16:32 crc kubenswrapper[4872]: I0126 09:16:32.363764 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" event={"ID":"68e8ae60-4d44-412d-be0a-575f475d5117","Type":"ContainerStarted","Data":"a3f3ed7d1988ca8560f73f3b4d89397697596d63d6d5905f640188abbb411f02"} Jan 26 09:16:32 crc kubenswrapper[4872]: I0126 09:16:32.363773 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" event={"ID":"68e8ae60-4d44-412d-be0a-575f475d5117","Type":"ContainerStarted","Data":"3886603eacdff55bbb0b88c8902c3c8f889a5e3ccabd167398c2347c2fc79777"} Jan 26 09:16:32 crc kubenswrapper[4872]: I0126 09:16:32.363786 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" event={"ID":"68e8ae60-4d44-412d-be0a-575f475d5117","Type":"ContainerStarted","Data":"1b159677e7139148e17475ac62231361556f5ba8cd4c92b351c256ff6d70a2e7"} Jan 26 09:16:33 crc kubenswrapper[4872]: I0126 09:16:33.196751 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2577331-1886-4abb-896b-8d8a112e40d8" path="/var/lib/kubelet/pods/e2577331-1886-4abb-896b-8d8a112e40d8/volumes" Jan 26 09:16:35 crc kubenswrapper[4872]: I0126 09:16:35.384601 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" event={"ID":"68e8ae60-4d44-412d-be0a-575f475d5117","Type":"ContainerStarted","Data":"0e9cffba4f0ff4ab8b932a2b20f94837f40e47f5a37971600bf24f015adb4ab9"} Jan 26 09:16:37 crc kubenswrapper[4872]: I0126 09:16:37.402912 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" event={"ID":"68e8ae60-4d44-412d-be0a-575f475d5117","Type":"ContainerStarted","Data":"20f9f8e53ee3eb36f08a24ac6c4f1689adc4ec9b82a30890ccbfa6c741b636ef"} Jan 26 09:16:37 crc kubenswrapper[4872]: I0126 09:16:37.403887 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:37 crc kubenswrapper[4872]: I0126 09:16:37.435430 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:37 crc kubenswrapper[4872]: I0126 09:16:37.443772 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" podStartSLOduration=7.443749749 podStartE2EDuration="7.443749749s" podCreationTimestamp="2026-01-26 09:16:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:16:37.438940945 +0000 UTC m=+530.747780766" watchObservedRunningTime="2026-01-26 09:16:37.443749749 +0000 UTC m=+530.752589550" Jan 26 09:16:38 crc kubenswrapper[4872]: I0126 09:16:38.409187 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:38 crc kubenswrapper[4872]: I0126 09:16:38.409255 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:38 crc kubenswrapper[4872]: I0126 09:16:38.438220 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:16:40 crc kubenswrapper[4872]: I0126 09:16:40.229247 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:16:40 crc kubenswrapper[4872]: I0126 09:16:40.229665 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:16:44 crc kubenswrapper[4872]: I0126 09:16:44.184321 4872 scope.go:117] "RemoveContainer" containerID="4960fa5dfa153921fd4c32dffe9662db8304b2a8242ef2b570ccc174c76279d8" Jan 26 09:16:44 crc kubenswrapper[4872]: I0126 09:16:44.443962 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-swcf5_3579107d-a172-4905-a491-6dd2b5a304a5/kube-multus/1.log" Jan 26 09:16:44 crc kubenswrapper[4872]: I0126 09:16:44.444037 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-swcf5" event={"ID":"3579107d-a172-4905-a491-6dd2b5a304a5","Type":"ContainerStarted","Data":"f697421f2773d135850e429e15829e23e30dbbdcdf1fa77a99a1e5e55ad10aa7"} Jan 26 09:17:00 crc kubenswrapper[4872]: I0126 09:17:00.887665 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cqtgt" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.046364 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb"] Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.047876 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.050965 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.058816 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb"] Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.148213 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/077c7b6a-93a7-4851-84fe-549a78533721-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb\" (UID: \"077c7b6a-93a7-4851-84fe-549a78533721\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.148290 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtwmk\" (UniqueName: \"kubernetes.io/projected/077c7b6a-93a7-4851-84fe-549a78533721-kube-api-access-vtwmk\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb\" (UID: \"077c7b6a-93a7-4851-84fe-549a78533721\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.148548 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/077c7b6a-93a7-4851-84fe-549a78533721-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb\" (UID: \"077c7b6a-93a7-4851-84fe-549a78533721\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.250489 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/077c7b6a-93a7-4851-84fe-549a78533721-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb\" (UID: \"077c7b6a-93a7-4851-84fe-549a78533721\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.250551 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtwmk\" (UniqueName: \"kubernetes.io/projected/077c7b6a-93a7-4851-84fe-549a78533721-kube-api-access-vtwmk\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb\" (UID: \"077c7b6a-93a7-4851-84fe-549a78533721\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.250599 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/077c7b6a-93a7-4851-84fe-549a78533721-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb\" (UID: \"077c7b6a-93a7-4851-84fe-549a78533721\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.251223 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/077c7b6a-93a7-4851-84fe-549a78533721-bundle\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb\" (UID: \"077c7b6a-93a7-4851-84fe-549a78533721\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.251288 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/077c7b6a-93a7-4851-84fe-549a78533721-util\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb\" (UID: \"077c7b6a-93a7-4851-84fe-549a78533721\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.270785 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtwmk\" (UniqueName: \"kubernetes.io/projected/077c7b6a-93a7-4851-84fe-549a78533721-kube-api-access-vtwmk\") pod \"53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb\" (UID: \"077c7b6a-93a7-4851-84fe-549a78533721\") " pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.371123 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:02 crc kubenswrapper[4872]: I0126 09:17:02.809757 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb"] Jan 26 09:17:03 crc kubenswrapper[4872]: I0126 09:17:03.563139 4872 generic.go:334] "Generic (PLEG): container finished" podID="077c7b6a-93a7-4851-84fe-549a78533721" containerID="e90d00dc5ffba1f10a027650de51d0083e1c70b942e95b8e8b4c932661faa099" exitCode=0 Jan 26 09:17:03 crc kubenswrapper[4872]: I0126 09:17:03.563337 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" event={"ID":"077c7b6a-93a7-4851-84fe-549a78533721","Type":"ContainerDied","Data":"e90d00dc5ffba1f10a027650de51d0083e1c70b942e95b8e8b4c932661faa099"} Jan 26 09:17:03 crc kubenswrapper[4872]: I0126 09:17:03.563603 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" event={"ID":"077c7b6a-93a7-4851-84fe-549a78533721","Type":"ContainerStarted","Data":"71cdd66d1d26dfc69f512be56eb009ba41813f0063c3eb82061b2071103271ff"} Jan 26 09:17:05 crc kubenswrapper[4872]: I0126 09:17:05.583466 4872 generic.go:334] "Generic (PLEG): container finished" podID="077c7b6a-93a7-4851-84fe-549a78533721" containerID="e71c82ea25bd99872b7b870315ce1d8066eed453d7396162987c584012eba520" exitCode=0 Jan 26 09:17:05 crc kubenswrapper[4872]: I0126 09:17:05.583562 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" event={"ID":"077c7b6a-93a7-4851-84fe-549a78533721","Type":"ContainerDied","Data":"e71c82ea25bd99872b7b870315ce1d8066eed453d7396162987c584012eba520"} Jan 26 09:17:06 crc kubenswrapper[4872]: I0126 09:17:06.591137 4872 generic.go:334] "Generic (PLEG): container finished" podID="077c7b6a-93a7-4851-84fe-549a78533721" containerID="236420723d3844bb4f1780118a6fa998500d27eb57f70a32793480ada1c499fb" exitCode=0 Jan 26 09:17:06 crc kubenswrapper[4872]: I0126 09:17:06.591235 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" event={"ID":"077c7b6a-93a7-4851-84fe-549a78533721","Type":"ContainerDied","Data":"236420723d3844bb4f1780118a6fa998500d27eb57f70a32793480ada1c499fb"} Jan 26 09:17:07 crc kubenswrapper[4872]: I0126 09:17:07.838966 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:07 crc kubenswrapper[4872]: I0126 09:17:07.931626 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtwmk\" (UniqueName: \"kubernetes.io/projected/077c7b6a-93a7-4851-84fe-549a78533721-kube-api-access-vtwmk\") pod \"077c7b6a-93a7-4851-84fe-549a78533721\" (UID: \"077c7b6a-93a7-4851-84fe-549a78533721\") " Jan 26 09:17:07 crc kubenswrapper[4872]: I0126 09:17:07.931916 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/077c7b6a-93a7-4851-84fe-549a78533721-util\") pod \"077c7b6a-93a7-4851-84fe-549a78533721\" (UID: \"077c7b6a-93a7-4851-84fe-549a78533721\") " Jan 26 09:17:07 crc kubenswrapper[4872]: I0126 09:17:07.931961 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/077c7b6a-93a7-4851-84fe-549a78533721-bundle\") pod \"077c7b6a-93a7-4851-84fe-549a78533721\" (UID: \"077c7b6a-93a7-4851-84fe-549a78533721\") " Jan 26 09:17:07 crc kubenswrapper[4872]: I0126 09:17:07.932708 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/077c7b6a-93a7-4851-84fe-549a78533721-bundle" (OuterVolumeSpecName: "bundle") pod "077c7b6a-93a7-4851-84fe-549a78533721" (UID: "077c7b6a-93a7-4851-84fe-549a78533721"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:17:07 crc kubenswrapper[4872]: I0126 09:17:07.937998 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/077c7b6a-93a7-4851-84fe-549a78533721-kube-api-access-vtwmk" (OuterVolumeSpecName: "kube-api-access-vtwmk") pod "077c7b6a-93a7-4851-84fe-549a78533721" (UID: "077c7b6a-93a7-4851-84fe-549a78533721"). InnerVolumeSpecName "kube-api-access-vtwmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:17:07 crc kubenswrapper[4872]: I0126 09:17:07.953842 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/077c7b6a-93a7-4851-84fe-549a78533721-util" (OuterVolumeSpecName: "util") pod "077c7b6a-93a7-4851-84fe-549a78533721" (UID: "077c7b6a-93a7-4851-84fe-549a78533721"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:17:08 crc kubenswrapper[4872]: I0126 09:17:08.033768 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtwmk\" (UniqueName: \"kubernetes.io/projected/077c7b6a-93a7-4851-84fe-549a78533721-kube-api-access-vtwmk\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:08 crc kubenswrapper[4872]: I0126 09:17:08.033866 4872 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/077c7b6a-93a7-4851-84fe-549a78533721-util\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:08 crc kubenswrapper[4872]: I0126 09:17:08.033876 4872 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/077c7b6a-93a7-4851-84fe-549a78533721-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:08 crc kubenswrapper[4872]: I0126 09:17:08.608018 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" event={"ID":"077c7b6a-93a7-4851-84fe-549a78533721","Type":"ContainerDied","Data":"71cdd66d1d26dfc69f512be56eb009ba41813f0063c3eb82061b2071103271ff"} Jan 26 09:17:08 crc kubenswrapper[4872]: I0126 09:17:08.608072 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71cdd66d1d26dfc69f512be56eb009ba41813f0063c3eb82061b2071103271ff" Jan 26 09:17:08 crc kubenswrapper[4872]: I0126 09:17:08.608087 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.228092 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.228164 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.624249 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-b72fb"] Jan 26 09:17:10 crc kubenswrapper[4872]: E0126 09:17:10.625015 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="077c7b6a-93a7-4851-84fe-549a78533721" containerName="util" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.625036 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="077c7b6a-93a7-4851-84fe-549a78533721" containerName="util" Jan 26 09:17:10 crc kubenswrapper[4872]: E0126 09:17:10.625058 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="077c7b6a-93a7-4851-84fe-549a78533721" containerName="extract" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.625069 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="077c7b6a-93a7-4851-84fe-549a78533721" containerName="extract" Jan 26 09:17:10 crc kubenswrapper[4872]: E0126 09:17:10.625098 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="077c7b6a-93a7-4851-84fe-549a78533721" containerName="pull" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.625104 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="077c7b6a-93a7-4851-84fe-549a78533721" containerName="pull" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.625218 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="077c7b6a-93a7-4851-84fe-549a78533721" containerName="extract" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.625650 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-b72fb" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.627991 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.628233 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-f4tnm" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.628538 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.637941 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-b72fb"] Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.773730 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w92qh\" (UniqueName: \"kubernetes.io/projected/d45d048b-dc90-4af0-800d-cdaa152c7634-kube-api-access-w92qh\") pod \"nmstate-operator-646758c888-b72fb\" (UID: \"d45d048b-dc90-4af0-800d-cdaa152c7634\") " pod="openshift-nmstate/nmstate-operator-646758c888-b72fb" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.875498 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w92qh\" (UniqueName: \"kubernetes.io/projected/d45d048b-dc90-4af0-800d-cdaa152c7634-kube-api-access-w92qh\") pod \"nmstate-operator-646758c888-b72fb\" (UID: \"d45d048b-dc90-4af0-800d-cdaa152c7634\") " pod="openshift-nmstate/nmstate-operator-646758c888-b72fb" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.900868 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w92qh\" (UniqueName: \"kubernetes.io/projected/d45d048b-dc90-4af0-800d-cdaa152c7634-kube-api-access-w92qh\") pod \"nmstate-operator-646758c888-b72fb\" (UID: \"d45d048b-dc90-4af0-800d-cdaa152c7634\") " pod="openshift-nmstate/nmstate-operator-646758c888-b72fb" Jan 26 09:17:10 crc kubenswrapper[4872]: I0126 09:17:10.942494 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-646758c888-b72fb" Jan 26 09:17:11 crc kubenswrapper[4872]: I0126 09:17:11.178926 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-646758c888-b72fb"] Jan 26 09:17:11 crc kubenswrapper[4872]: I0126 09:17:11.625446 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-b72fb" event={"ID":"d45d048b-dc90-4af0-800d-cdaa152c7634","Type":"ContainerStarted","Data":"aa245001fb2034021a6dcb9e02512912b5a62649c712250dc328c049fa230237"} Jan 26 09:17:13 crc kubenswrapper[4872]: I0126 09:17:13.644662 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-646758c888-b72fb" event={"ID":"d45d048b-dc90-4af0-800d-cdaa152c7634","Type":"ContainerStarted","Data":"17995fe7e481d63aba5314bfb6b4ff74266c3abc9aa78920c8095f399c85e85f"} Jan 26 09:17:13 crc kubenswrapper[4872]: I0126 09:17:13.667545 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-646758c888-b72fb" podStartSLOduration=1.5911096850000002 podStartE2EDuration="3.667513425s" podCreationTimestamp="2026-01-26 09:17:10 +0000 UTC" firstStartedPulling="2026-01-26 09:17:11.186420642 +0000 UTC m=+564.495260443" lastFinishedPulling="2026-01-26 09:17:13.262824382 +0000 UTC m=+566.571664183" observedRunningTime="2026-01-26 09:17:13.663305047 +0000 UTC m=+566.972144858" watchObservedRunningTime="2026-01-26 09:17:13.667513425 +0000 UTC m=+566.976353246" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.561470 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-rzhxm"] Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.562950 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-rzhxm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.567564 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-hs9fl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.573343 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-rzhxm"] Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.587045 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv"] Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.588009 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.591866 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.636809 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-cbbnl"] Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.640078 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.674718 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv"] Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.722740 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm"] Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.723828 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.731201 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm"] Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.732000 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d7ecce6c-87f7-4b85-aceb-1430b1668e0d-nmstate-lock\") pod \"nmstate-handler-cbbnl\" (UID: \"d7ecce6c-87f7-4b85-aceb-1430b1668e0d\") " pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.732069 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vms6\" (UniqueName: \"kubernetes.io/projected/464e800f-b945-4d14-8fb6-d023d82ac993-kube-api-access-9vms6\") pod \"nmstate-metrics-54757c584b-rzhxm\" (UID: \"464e800f-b945-4d14-8fb6-d023d82ac993\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-rzhxm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.732108 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d7ecce6c-87f7-4b85-aceb-1430b1668e0d-ovs-socket\") pod \"nmstate-handler-cbbnl\" (UID: \"d7ecce6c-87f7-4b85-aceb-1430b1668e0d\") " pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.732137 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d7ecce6c-87f7-4b85-aceb-1430b1668e0d-dbus-socket\") pod \"nmstate-handler-cbbnl\" (UID: \"d7ecce6c-87f7-4b85-aceb-1430b1668e0d\") " pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.732184 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2z5t\" (UniqueName: \"kubernetes.io/projected/3020f909-4524-4ae9-9fcb-5194b45565de-kube-api-access-q2z5t\") pod \"nmstate-webhook-8474b5b9d8-8cgjv\" (UID: \"3020f909-4524-4ae9-9fcb-5194b45565de\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.732219 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pplng\" (UniqueName: \"kubernetes.io/projected/d7ecce6c-87f7-4b85-aceb-1430b1668e0d-kube-api-access-pplng\") pod \"nmstate-handler-cbbnl\" (UID: \"d7ecce6c-87f7-4b85-aceb-1430b1668e0d\") " pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.732249 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3020f909-4524-4ae9-9fcb-5194b45565de-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-8cgjv\" (UID: \"3020f909-4524-4ae9-9fcb-5194b45565de\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.733382 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-j4nvq" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.733545 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.733818 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.833999 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pplng\" (UniqueName: \"kubernetes.io/projected/d7ecce6c-87f7-4b85-aceb-1430b1668e0d-kube-api-access-pplng\") pod \"nmstate-handler-cbbnl\" (UID: \"d7ecce6c-87f7-4b85-aceb-1430b1668e0d\") " pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.834450 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3020f909-4524-4ae9-9fcb-5194b45565de-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-8cgjv\" (UID: \"3020f909-4524-4ae9-9fcb-5194b45565de\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.834492 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d7ecce6c-87f7-4b85-aceb-1430b1668e0d-nmstate-lock\") pod \"nmstate-handler-cbbnl\" (UID: \"d7ecce6c-87f7-4b85-aceb-1430b1668e0d\") " pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.834535 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vms6\" (UniqueName: \"kubernetes.io/projected/464e800f-b945-4d14-8fb6-d023d82ac993-kube-api-access-9vms6\") pod \"nmstate-metrics-54757c584b-rzhxm\" (UID: \"464e800f-b945-4d14-8fb6-d023d82ac993\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-rzhxm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.834574 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fead2890-af05-404f-a7d4-6321b48e31a2-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-lk6cm\" (UID: \"fead2890-af05-404f-a7d4-6321b48e31a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.834602 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fead2890-af05-404f-a7d4-6321b48e31a2-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-lk6cm\" (UID: \"fead2890-af05-404f-a7d4-6321b48e31a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.834622 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d7ecce6c-87f7-4b85-aceb-1430b1668e0d-ovs-socket\") pod \"nmstate-handler-cbbnl\" (UID: \"d7ecce6c-87f7-4b85-aceb-1430b1668e0d\") " pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.834649 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d7ecce6c-87f7-4b85-aceb-1430b1668e0d-dbus-socket\") pod \"nmstate-handler-cbbnl\" (UID: \"d7ecce6c-87f7-4b85-aceb-1430b1668e0d\") " pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.834655 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/d7ecce6c-87f7-4b85-aceb-1430b1668e0d-nmstate-lock\") pod \"nmstate-handler-cbbnl\" (UID: \"d7ecce6c-87f7-4b85-aceb-1430b1668e0d\") " pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.834684 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxc97\" (UniqueName: \"kubernetes.io/projected/fead2890-af05-404f-a7d4-6321b48e31a2-kube-api-access-gxc97\") pod \"nmstate-console-plugin-7754f76f8b-lk6cm\" (UID: \"fead2890-af05-404f-a7d4-6321b48e31a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.834746 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/d7ecce6c-87f7-4b85-aceb-1430b1668e0d-ovs-socket\") pod \"nmstate-handler-cbbnl\" (UID: \"d7ecce6c-87f7-4b85-aceb-1430b1668e0d\") " pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.834912 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2z5t\" (UniqueName: \"kubernetes.io/projected/3020f909-4524-4ae9-9fcb-5194b45565de-kube-api-access-q2z5t\") pod \"nmstate-webhook-8474b5b9d8-8cgjv\" (UID: \"3020f909-4524-4ae9-9fcb-5194b45565de\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.835001 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/d7ecce6c-87f7-4b85-aceb-1430b1668e0d-dbus-socket\") pod \"nmstate-handler-cbbnl\" (UID: \"d7ecce6c-87f7-4b85-aceb-1430b1668e0d\") " pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.843607 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/3020f909-4524-4ae9-9fcb-5194b45565de-tls-key-pair\") pod \"nmstate-webhook-8474b5b9d8-8cgjv\" (UID: \"3020f909-4524-4ae9-9fcb-5194b45565de\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.856747 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vms6\" (UniqueName: \"kubernetes.io/projected/464e800f-b945-4d14-8fb6-d023d82ac993-kube-api-access-9vms6\") pod \"nmstate-metrics-54757c584b-rzhxm\" (UID: \"464e800f-b945-4d14-8fb6-d023d82ac993\") " pod="openshift-nmstate/nmstate-metrics-54757c584b-rzhxm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.859560 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2z5t\" (UniqueName: \"kubernetes.io/projected/3020f909-4524-4ae9-9fcb-5194b45565de-kube-api-access-q2z5t\") pod \"nmstate-webhook-8474b5b9d8-8cgjv\" (UID: \"3020f909-4524-4ae9-9fcb-5194b45565de\") " pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.859631 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pplng\" (UniqueName: \"kubernetes.io/projected/d7ecce6c-87f7-4b85-aceb-1430b1668e0d-kube-api-access-pplng\") pod \"nmstate-handler-cbbnl\" (UID: \"d7ecce6c-87f7-4b85-aceb-1430b1668e0d\") " pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.887867 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-54757c584b-rzhxm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.905613 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.936041 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fead2890-af05-404f-a7d4-6321b48e31a2-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-lk6cm\" (UID: \"fead2890-af05-404f-a7d4-6321b48e31a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.936086 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fead2890-af05-404f-a7d4-6321b48e31a2-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-lk6cm\" (UID: \"fead2890-af05-404f-a7d4-6321b48e31a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.936129 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxc97\" (UniqueName: \"kubernetes.io/projected/fead2890-af05-404f-a7d4-6321b48e31a2-kube-api-access-gxc97\") pod \"nmstate-console-plugin-7754f76f8b-lk6cm\" (UID: \"fead2890-af05-404f-a7d4-6321b48e31a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.937770 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/fead2890-af05-404f-a7d4-6321b48e31a2-nginx-conf\") pod \"nmstate-console-plugin-7754f76f8b-lk6cm\" (UID: \"fead2890-af05-404f-a7d4-6321b48e31a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.951021 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/fead2890-af05-404f-a7d4-6321b48e31a2-plugin-serving-cert\") pod \"nmstate-console-plugin-7754f76f8b-lk6cm\" (UID: \"fead2890-af05-404f-a7d4-6321b48e31a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.965482 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxc97\" (UniqueName: \"kubernetes.io/projected/fead2890-af05-404f-a7d4-6321b48e31a2-kube-api-access-gxc97\") pod \"nmstate-console-plugin-7754f76f8b-lk6cm\" (UID: \"fead2890-af05-404f-a7d4-6321b48e31a2\") " pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.974637 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.975224 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6cb59dc76d-bmjqv"] Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.976234 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:14 crc kubenswrapper[4872]: I0126 09:17:14.979866 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6cb59dc76d-bmjqv"] Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.050103 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.141937 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/81995d5c-89fb-40af-af3e-99c4ee2de5e2-console-config\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.142010 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/81995d5c-89fb-40af-af3e-99c4ee2de5e2-console-serving-cert\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.142049 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/81995d5c-89fb-40af-af3e-99c4ee2de5e2-oauth-serving-cert\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.142072 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81995d5c-89fb-40af-af3e-99c4ee2de5e2-trusted-ca-bundle\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.142097 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggd64\" (UniqueName: \"kubernetes.io/projected/81995d5c-89fb-40af-af3e-99c4ee2de5e2-kube-api-access-ggd64\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.142126 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/81995d5c-89fb-40af-af3e-99c4ee2de5e2-console-oauth-config\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.142145 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/81995d5c-89fb-40af-af3e-99c4ee2de5e2-service-ca\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.219929 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv"] Jan 26 09:17:15 crc kubenswrapper[4872]: W0126 09:17:15.225814 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3020f909_4524_4ae9_9fcb_5194b45565de.slice/crio-d3ff022505a1d760945827fd09bf28571b09e91c6b69daa452580623bb0edc74 WatchSource:0}: Error finding container d3ff022505a1d760945827fd09bf28571b09e91c6b69daa452580623bb0edc74: Status 404 returned error can't find the container with id d3ff022505a1d760945827fd09bf28571b09e91c6b69daa452580623bb0edc74 Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.244014 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81995d5c-89fb-40af-af3e-99c4ee2de5e2-trusted-ca-bundle\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.244066 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/81995d5c-89fb-40af-af3e-99c4ee2de5e2-oauth-serving-cert\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.244112 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggd64\" (UniqueName: \"kubernetes.io/projected/81995d5c-89fb-40af-af3e-99c4ee2de5e2-kube-api-access-ggd64\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.244156 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/81995d5c-89fb-40af-af3e-99c4ee2de5e2-console-oauth-config\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.244183 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/81995d5c-89fb-40af-af3e-99c4ee2de5e2-service-ca\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.244263 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/81995d5c-89fb-40af-af3e-99c4ee2de5e2-console-config\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.244304 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/81995d5c-89fb-40af-af3e-99c4ee2de5e2-console-serving-cert\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.245460 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/81995d5c-89fb-40af-af3e-99c4ee2de5e2-console-config\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.245515 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/81995d5c-89fb-40af-af3e-99c4ee2de5e2-oauth-serving-cert\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.245960 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/81995d5c-89fb-40af-af3e-99c4ee2de5e2-service-ca\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.249710 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81995d5c-89fb-40af-af3e-99c4ee2de5e2-trusted-ca-bundle\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.249991 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/81995d5c-89fb-40af-af3e-99c4ee2de5e2-console-oauth-config\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.250306 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/81995d5c-89fb-40af-af3e-99c4ee2de5e2-console-serving-cert\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.262376 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggd64\" (UniqueName: \"kubernetes.io/projected/81995d5c-89fb-40af-af3e-99c4ee2de5e2-kube-api-access-ggd64\") pod \"console-6cb59dc76d-bmjqv\" (UID: \"81995d5c-89fb-40af-af3e-99c4ee2de5e2\") " pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.298213 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.304045 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm"] Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.374960 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-54757c584b-rzhxm"] Jan 26 09:17:15 crc kubenswrapper[4872]: W0126 09:17:15.400210 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod464e800f_b945_4d14_8fb6_d023d82ac993.slice/crio-61554898e59386b07b44c2f28dcf0f8dedd2fed787be0dfd1d1cde9f765a6543 WatchSource:0}: Error finding container 61554898e59386b07b44c2f28dcf0f8dedd2fed787be0dfd1d1cde9f765a6543: Status 404 returned error can't find the container with id 61554898e59386b07b44c2f28dcf0f8dedd2fed787be0dfd1d1cde9f765a6543 Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.541512 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6cb59dc76d-bmjqv"] Jan 26 09:17:15 crc kubenswrapper[4872]: W0126 09:17:15.544627 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81995d5c_89fb_40af_af3e_99c4ee2de5e2.slice/crio-b00c6e0e6cf15ff789e99acad3d82b2565b6e4a5d58350ca662da3aeffaa7d9e WatchSource:0}: Error finding container b00c6e0e6cf15ff789e99acad3d82b2565b6e4a5d58350ca662da3aeffaa7d9e: Status 404 returned error can't find the container with id b00c6e0e6cf15ff789e99acad3d82b2565b6e4a5d58350ca662da3aeffaa7d9e Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.662134 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-cbbnl" event={"ID":"d7ecce6c-87f7-4b85-aceb-1430b1668e0d","Type":"ContainerStarted","Data":"00648307d0b69413dce63b5012181cf8ca99bcd6ff3c566377450ad3b02faafc"} Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.663652 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" event={"ID":"fead2890-af05-404f-a7d4-6321b48e31a2","Type":"ContainerStarted","Data":"2db46d1b7ed6f4d6aa8760c98319362919fd914b4fff99408b22acc012add85a"} Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.664866 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6cb59dc76d-bmjqv" event={"ID":"81995d5c-89fb-40af-af3e-99c4ee2de5e2","Type":"ContainerStarted","Data":"b00c6e0e6cf15ff789e99acad3d82b2565b6e4a5d58350ca662da3aeffaa7d9e"} Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.665842 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-rzhxm" event={"ID":"464e800f-b945-4d14-8fb6-d023d82ac993","Type":"ContainerStarted","Data":"61554898e59386b07b44c2f28dcf0f8dedd2fed787be0dfd1d1cde9f765a6543"} Jan 26 09:17:15 crc kubenswrapper[4872]: I0126 09:17:15.666699 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" event={"ID":"3020f909-4524-4ae9-9fcb-5194b45565de","Type":"ContainerStarted","Data":"d3ff022505a1d760945827fd09bf28571b09e91c6b69daa452580623bb0edc74"} Jan 26 09:17:16 crc kubenswrapper[4872]: I0126 09:17:16.681396 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6cb59dc76d-bmjqv" event={"ID":"81995d5c-89fb-40af-af3e-99c4ee2de5e2","Type":"ContainerStarted","Data":"2c0d58124de746610ab5397bdfd25ebefd77c9a7b7ff523a455325eb70b8572a"} Jan 26 09:17:16 crc kubenswrapper[4872]: I0126 09:17:16.721937 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6cb59dc76d-bmjqv" podStartSLOduration=2.721910567 podStartE2EDuration="2.721910567s" podCreationTimestamp="2026-01-26 09:17:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:17:16.708358018 +0000 UTC m=+570.017197819" watchObservedRunningTime="2026-01-26 09:17:16.721910567 +0000 UTC m=+570.030750368" Jan 26 09:17:18 crc kubenswrapper[4872]: I0126 09:17:18.696426 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-cbbnl" event={"ID":"d7ecce6c-87f7-4b85-aceb-1430b1668e0d","Type":"ContainerStarted","Data":"4908c21818c876be5692fbd8ddf5e67626e38fc0e74a0cc571352b9c19a76b4e"} Jan 26 09:17:18 crc kubenswrapper[4872]: I0126 09:17:18.697000 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:18 crc kubenswrapper[4872]: I0126 09:17:18.700043 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" event={"ID":"3020f909-4524-4ae9-9fcb-5194b45565de","Type":"ContainerStarted","Data":"ce40eeef1bd09b097d3ba4bf7bb4f215c417b9df6dd3cb7e11e5ff4ab3df993a"} Jan 26 09:17:18 crc kubenswrapper[4872]: I0126 09:17:18.700218 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" Jan 26 09:17:18 crc kubenswrapper[4872]: I0126 09:17:18.702202 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" event={"ID":"fead2890-af05-404f-a7d4-6321b48e31a2","Type":"ContainerStarted","Data":"c2cb1fda641364cef04834a856bdf998aa09c0b23660041120dc752484bd3736"} Jan 26 09:17:18 crc kubenswrapper[4872]: I0126 09:17:18.706366 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-rzhxm" event={"ID":"464e800f-b945-4d14-8fb6-d023d82ac993","Type":"ContainerStarted","Data":"9fb2cb716d45085c2c8d41992c34224a513022e3d306bafd352b986cbcd72049"} Jan 26 09:17:18 crc kubenswrapper[4872]: I0126 09:17:18.721844 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-cbbnl" podStartSLOduration=1.4262104469999999 podStartE2EDuration="4.72182709s" podCreationTimestamp="2026-01-26 09:17:14 +0000 UTC" firstStartedPulling="2026-01-26 09:17:15.015461608 +0000 UTC m=+568.324301409" lastFinishedPulling="2026-01-26 09:17:18.311078251 +0000 UTC m=+571.619918052" observedRunningTime="2026-01-26 09:17:18.71679078 +0000 UTC m=+572.025630591" watchObservedRunningTime="2026-01-26 09:17:18.72182709 +0000 UTC m=+572.030666891" Jan 26 09:17:18 crc kubenswrapper[4872]: I0126 09:17:18.738604 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" podStartSLOduration=1.695476521 podStartE2EDuration="4.738585051s" podCreationTimestamp="2026-01-26 09:17:14 +0000 UTC" firstStartedPulling="2026-01-26 09:17:15.229141602 +0000 UTC m=+568.537981393" lastFinishedPulling="2026-01-26 09:17:18.272250102 +0000 UTC m=+571.581089923" observedRunningTime="2026-01-26 09:17:18.736137297 +0000 UTC m=+572.044977108" watchObservedRunningTime="2026-01-26 09:17:18.738585051 +0000 UTC m=+572.047424842" Jan 26 09:17:18 crc kubenswrapper[4872]: I0126 09:17:18.760859 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7754f76f8b-lk6cm" podStartSLOduration=1.809530952 podStartE2EDuration="4.760833903s" podCreationTimestamp="2026-01-26 09:17:14 +0000 UTC" firstStartedPulling="2026-01-26 09:17:15.321026593 +0000 UTC m=+568.629866394" lastFinishedPulling="2026-01-26 09:17:18.272329544 +0000 UTC m=+571.581169345" observedRunningTime="2026-01-26 09:17:18.75684537 +0000 UTC m=+572.065685171" watchObservedRunningTime="2026-01-26 09:17:18.760833903 +0000 UTC m=+572.069673704" Jan 26 09:17:21 crc kubenswrapper[4872]: I0126 09:17:21.729284 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-54757c584b-rzhxm" event={"ID":"464e800f-b945-4d14-8fb6-d023d82ac993","Type":"ContainerStarted","Data":"58e89c17ea4c56cd3026b1b976a44f92009558d5a4d27ce85ca38935a8b2a781"} Jan 26 09:17:21 crc kubenswrapper[4872]: I0126 09:17:21.760119 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-54757c584b-rzhxm" podStartSLOduration=2.065442741 podStartE2EDuration="7.760088717s" podCreationTimestamp="2026-01-26 09:17:14 +0000 UTC" firstStartedPulling="2026-01-26 09:17:15.410318859 +0000 UTC m=+568.719158660" lastFinishedPulling="2026-01-26 09:17:21.104964825 +0000 UTC m=+574.413804636" observedRunningTime="2026-01-26 09:17:21.751096036 +0000 UTC m=+575.059935877" watchObservedRunningTime="2026-01-26 09:17:21.760088717 +0000 UTC m=+575.068928738" Jan 26 09:17:25 crc kubenswrapper[4872]: I0126 09:17:25.004350 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-cbbnl" Jan 26 09:17:25 crc kubenswrapper[4872]: I0126 09:17:25.298989 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:25 crc kubenswrapper[4872]: I0126 09:17:25.299066 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:25 crc kubenswrapper[4872]: I0126 09:17:25.305209 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:25 crc kubenswrapper[4872]: I0126 09:17:25.764721 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6cb59dc76d-bmjqv" Jan 26 09:17:25 crc kubenswrapper[4872]: I0126 09:17:25.826361 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ztd4t"] Jan 26 09:17:34 crc kubenswrapper[4872]: I0126 09:17:34.926034 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-8474b5b9d8-8cgjv" Jan 26 09:17:40 crc kubenswrapper[4872]: I0126 09:17:40.227632 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:17:40 crc kubenswrapper[4872]: I0126 09:17:40.228288 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:17:40 crc kubenswrapper[4872]: I0126 09:17:40.228371 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:17:40 crc kubenswrapper[4872]: I0126 09:17:40.229427 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b2ddad1c18cf250920b95ba098a58a140582a8cf9f6b8c8479e40b6449d7e085"} pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 09:17:40 crc kubenswrapper[4872]: I0126 09:17:40.229495 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" containerID="cri-o://b2ddad1c18cf250920b95ba098a58a140582a8cf9f6b8c8479e40b6449d7e085" gracePeriod=600 Jan 26 09:17:40 crc kubenswrapper[4872]: I0126 09:17:40.863565 4872 generic.go:334] "Generic (PLEG): container finished" podID="fca44d96-a000-4bf2-8283-a937b0192880" containerID="b2ddad1c18cf250920b95ba098a58a140582a8cf9f6b8c8479e40b6449d7e085" exitCode=0 Jan 26 09:17:40 crc kubenswrapper[4872]: I0126 09:17:40.863657 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerDied","Data":"b2ddad1c18cf250920b95ba098a58a140582a8cf9f6b8c8479e40b6449d7e085"} Jan 26 09:17:40 crc kubenswrapper[4872]: I0126 09:17:40.864100 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerStarted","Data":"df957db96f1d4f96c1bc63715f9b57ffc3271e0d6dac35d880b93d7745dd43b0"} Jan 26 09:17:40 crc kubenswrapper[4872]: I0126 09:17:40.864130 4872 scope.go:117] "RemoveContainer" containerID="fc57e5384dec6c48143f4ceff8720fb743fb9d09433bb2bd3b7ba7d5cc00b27c" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.012112 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68"] Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.016458 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.023096 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.026237 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68"] Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.081012 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x6gv\" (UniqueName: \"kubernetes.io/projected/18a3643b-548e-4f24-9489-7271c5d5dfee-kube-api-access-6x6gv\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68\" (UID: \"18a3643b-548e-4f24-9489-7271c5d5dfee\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.081072 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/18a3643b-548e-4f24-9489-7271c5d5dfee-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68\" (UID: \"18a3643b-548e-4f24-9489-7271c5d5dfee\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.081116 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/18a3643b-548e-4f24-9489-7271c5d5dfee-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68\" (UID: \"18a3643b-548e-4f24-9489-7271c5d5dfee\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.183069 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x6gv\" (UniqueName: \"kubernetes.io/projected/18a3643b-548e-4f24-9489-7271c5d5dfee-kube-api-access-6x6gv\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68\" (UID: \"18a3643b-548e-4f24-9489-7271c5d5dfee\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.183135 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/18a3643b-548e-4f24-9489-7271c5d5dfee-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68\" (UID: \"18a3643b-548e-4f24-9489-7271c5d5dfee\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.183186 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/18a3643b-548e-4f24-9489-7271c5d5dfee-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68\" (UID: \"18a3643b-548e-4f24-9489-7271c5d5dfee\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.183836 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/18a3643b-548e-4f24-9489-7271c5d5dfee-util\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68\" (UID: \"18a3643b-548e-4f24-9489-7271c5d5dfee\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.183969 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/18a3643b-548e-4f24-9489-7271c5d5dfee-bundle\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68\" (UID: \"18a3643b-548e-4f24-9489-7271c5d5dfee\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.213976 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x6gv\" (UniqueName: \"kubernetes.io/projected/18a3643b-548e-4f24-9489-7271c5d5dfee-kube-api-access-6x6gv\") pod \"270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68\" (UID: \"18a3643b-548e-4f24-9489-7271c5d5dfee\") " pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.335056 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.616228 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68"] Jan 26 09:17:48 crc kubenswrapper[4872]: W0126 09:17:48.627282 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18a3643b_548e_4f24_9489_7271c5d5dfee.slice/crio-044f026b683ec5bb94ede772a08cf63539b85c76b177949c5d1caa97b9c58fc6 WatchSource:0}: Error finding container 044f026b683ec5bb94ede772a08cf63539b85c76b177949c5d1caa97b9c58fc6: Status 404 returned error can't find the container with id 044f026b683ec5bb94ede772a08cf63539b85c76b177949c5d1caa97b9c58fc6 Jan 26 09:17:48 crc kubenswrapper[4872]: I0126 09:17:48.934010 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" event={"ID":"18a3643b-548e-4f24-9489-7271c5d5dfee","Type":"ContainerStarted","Data":"044f026b683ec5bb94ede772a08cf63539b85c76b177949c5d1caa97b9c58fc6"} Jan 26 09:17:49 crc kubenswrapper[4872]: I0126 09:17:49.946709 4872 generic.go:334] "Generic (PLEG): container finished" podID="18a3643b-548e-4f24-9489-7271c5d5dfee" containerID="e32db6805c2ae7e4e2c5f2e08e30c747e28ce67a7a8e008361edc8cca23eb713" exitCode=0 Jan 26 09:17:49 crc kubenswrapper[4872]: I0126 09:17:49.946831 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" event={"ID":"18a3643b-548e-4f24-9489-7271c5d5dfee","Type":"ContainerDied","Data":"e32db6805c2ae7e4e2c5f2e08e30c747e28ce67a7a8e008361edc8cca23eb713"} Jan 26 09:17:50 crc kubenswrapper[4872]: I0126 09:17:50.878454 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-ztd4t" podUID="9d625f91-d4de-4687-bdbe-8ff985d7e406" containerName="console" containerID="cri-o://cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723" gracePeriod=15 Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.383996 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ztd4t_9d625f91-d4de-4687-bdbe-8ff985d7e406/console/0.log" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.384483 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.534031 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-config\") pod \"9d625f91-d4de-4687-bdbe-8ff985d7e406\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.534137 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-serving-cert\") pod \"9d625f91-d4de-4687-bdbe-8ff985d7e406\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.534652 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-oauth-serving-cert\") pod \"9d625f91-d4de-4687-bdbe-8ff985d7e406\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.534694 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-trusted-ca-bundle\") pod \"9d625f91-d4de-4687-bdbe-8ff985d7e406\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.534833 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-service-ca\") pod \"9d625f91-d4de-4687-bdbe-8ff985d7e406\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.534874 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rchtx\" (UniqueName: \"kubernetes.io/projected/9d625f91-d4de-4687-bdbe-8ff985d7e406-kube-api-access-rchtx\") pod \"9d625f91-d4de-4687-bdbe-8ff985d7e406\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.534917 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-oauth-config\") pod \"9d625f91-d4de-4687-bdbe-8ff985d7e406\" (UID: \"9d625f91-d4de-4687-bdbe-8ff985d7e406\") " Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.535593 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-config" (OuterVolumeSpecName: "console-config") pod "9d625f91-d4de-4687-bdbe-8ff985d7e406" (UID: "9d625f91-d4de-4687-bdbe-8ff985d7e406"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.535610 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-service-ca" (OuterVolumeSpecName: "service-ca") pod "9d625f91-d4de-4687-bdbe-8ff985d7e406" (UID: "9d625f91-d4de-4687-bdbe-8ff985d7e406"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.535692 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9d625f91-d4de-4687-bdbe-8ff985d7e406" (UID: "9d625f91-d4de-4687-bdbe-8ff985d7e406"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.536304 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9d625f91-d4de-4687-bdbe-8ff985d7e406" (UID: "9d625f91-d4de-4687-bdbe-8ff985d7e406"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.541731 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9d625f91-d4de-4687-bdbe-8ff985d7e406" (UID: "9d625f91-d4de-4687-bdbe-8ff985d7e406"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.541723 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d625f91-d4de-4687-bdbe-8ff985d7e406-kube-api-access-rchtx" (OuterVolumeSpecName: "kube-api-access-rchtx") pod "9d625f91-d4de-4687-bdbe-8ff985d7e406" (UID: "9d625f91-d4de-4687-bdbe-8ff985d7e406"). InnerVolumeSpecName "kube-api-access-rchtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.541899 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9d625f91-d4de-4687-bdbe-8ff985d7e406" (UID: "9d625f91-d4de-4687-bdbe-8ff985d7e406"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.636509 4872 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-service-ca\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.636553 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rchtx\" (UniqueName: \"kubernetes.io/projected/9d625f91-d4de-4687-bdbe-8ff985d7e406-kube-api-access-rchtx\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.636566 4872 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-oauth-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.636577 4872 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.636589 4872 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d625f91-d4de-4687-bdbe-8ff985d7e406-console-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.636600 4872 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.636614 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d625f91-d4de-4687-bdbe-8ff985d7e406-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.964505 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-ztd4t_9d625f91-d4de-4687-bdbe-8ff985d7e406/console/0.log" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.965087 4872 generic.go:334] "Generic (PLEG): container finished" podID="9d625f91-d4de-4687-bdbe-8ff985d7e406" containerID="cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723" exitCode=2 Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.965160 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ztd4t" event={"ID":"9d625f91-d4de-4687-bdbe-8ff985d7e406","Type":"ContainerDied","Data":"cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723"} Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.965197 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ztd4t" event={"ID":"9d625f91-d4de-4687-bdbe-8ff985d7e406","Type":"ContainerDied","Data":"346c36eba54b6a1cd4c7bbc2a46dbe0dfc15407ec6c13f11c6e4f54804ea3a1c"} Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.965221 4872 scope.go:117] "RemoveContainer" containerID="cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.965249 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ztd4t" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.968319 4872 generic.go:334] "Generic (PLEG): container finished" podID="18a3643b-548e-4f24-9489-7271c5d5dfee" containerID="a2c58cafc0ddd223d9e840b025a904895c178eb6421deb27510aeb1c8038a94b" exitCode=0 Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.968393 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" event={"ID":"18a3643b-548e-4f24-9489-7271c5d5dfee","Type":"ContainerDied","Data":"a2c58cafc0ddd223d9e840b025a904895c178eb6421deb27510aeb1c8038a94b"} Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.987578 4872 scope.go:117] "RemoveContainer" containerID="cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723" Jan 26 09:17:51 crc kubenswrapper[4872]: E0126 09:17:51.988361 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723\": container with ID starting with cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723 not found: ID does not exist" containerID="cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723" Jan 26 09:17:51 crc kubenswrapper[4872]: I0126 09:17:51.988418 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723"} err="failed to get container status \"cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723\": rpc error: code = NotFound desc = could not find container \"cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723\": container with ID starting with cde5b3b3b5e1163ce47ed32bd07a281920039bd833073447bf8644dd35dcb723 not found: ID does not exist" Jan 26 09:17:52 crc kubenswrapper[4872]: I0126 09:17:52.038042 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-ztd4t"] Jan 26 09:17:52 crc kubenswrapper[4872]: I0126 09:17:52.050439 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-ztd4t"] Jan 26 09:17:52 crc kubenswrapper[4872]: I0126 09:17:52.979155 4872 generic.go:334] "Generic (PLEG): container finished" podID="18a3643b-548e-4f24-9489-7271c5d5dfee" containerID="07285d7c71db1f691e9f6b68d3ebbfa636c442479b9c0517df5bd3064e9aaf38" exitCode=0 Jan 26 09:17:52 crc kubenswrapper[4872]: I0126 09:17:52.979212 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" event={"ID":"18a3643b-548e-4f24-9489-7271c5d5dfee","Type":"ContainerDied","Data":"07285d7c71db1f691e9f6b68d3ebbfa636c442479b9c0517df5bd3064e9aaf38"} Jan 26 09:17:53 crc kubenswrapper[4872]: I0126 09:17:53.193444 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d625f91-d4de-4687-bdbe-8ff985d7e406" path="/var/lib/kubelet/pods/9d625f91-d4de-4687-bdbe-8ff985d7e406/volumes" Jan 26 09:17:54 crc kubenswrapper[4872]: I0126 09:17:54.334858 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:17:54 crc kubenswrapper[4872]: I0126 09:17:54.483674 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/18a3643b-548e-4f24-9489-7271c5d5dfee-bundle\") pod \"18a3643b-548e-4f24-9489-7271c5d5dfee\" (UID: \"18a3643b-548e-4f24-9489-7271c5d5dfee\") " Jan 26 09:17:54 crc kubenswrapper[4872]: I0126 09:17:54.483755 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/18a3643b-548e-4f24-9489-7271c5d5dfee-util\") pod \"18a3643b-548e-4f24-9489-7271c5d5dfee\" (UID: \"18a3643b-548e-4f24-9489-7271c5d5dfee\") " Jan 26 09:17:54 crc kubenswrapper[4872]: I0126 09:17:54.483842 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6x6gv\" (UniqueName: \"kubernetes.io/projected/18a3643b-548e-4f24-9489-7271c5d5dfee-kube-api-access-6x6gv\") pod \"18a3643b-548e-4f24-9489-7271c5d5dfee\" (UID: \"18a3643b-548e-4f24-9489-7271c5d5dfee\") " Jan 26 09:17:54 crc kubenswrapper[4872]: I0126 09:17:54.485701 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a3643b-548e-4f24-9489-7271c5d5dfee-bundle" (OuterVolumeSpecName: "bundle") pod "18a3643b-548e-4f24-9489-7271c5d5dfee" (UID: "18a3643b-548e-4f24-9489-7271c5d5dfee"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:17:54 crc kubenswrapper[4872]: I0126 09:17:54.492144 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18a3643b-548e-4f24-9489-7271c5d5dfee-kube-api-access-6x6gv" (OuterVolumeSpecName: "kube-api-access-6x6gv") pod "18a3643b-548e-4f24-9489-7271c5d5dfee" (UID: "18a3643b-548e-4f24-9489-7271c5d5dfee"). InnerVolumeSpecName "kube-api-access-6x6gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:17:54 crc kubenswrapper[4872]: I0126 09:17:54.501235 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a3643b-548e-4f24-9489-7271c5d5dfee-util" (OuterVolumeSpecName: "util") pod "18a3643b-548e-4f24-9489-7271c5d5dfee" (UID: "18a3643b-548e-4f24-9489-7271c5d5dfee"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:17:54 crc kubenswrapper[4872]: I0126 09:17:54.585579 4872 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/18a3643b-548e-4f24-9489-7271c5d5dfee-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:54 crc kubenswrapper[4872]: I0126 09:17:54.585637 4872 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/18a3643b-548e-4f24-9489-7271c5d5dfee-util\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:54 crc kubenswrapper[4872]: I0126 09:17:54.585650 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6x6gv\" (UniqueName: \"kubernetes.io/projected/18a3643b-548e-4f24-9489-7271c5d5dfee-kube-api-access-6x6gv\") on node \"crc\" DevicePath \"\"" Jan 26 09:17:55 crc kubenswrapper[4872]: I0126 09:17:55.006004 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" event={"ID":"18a3643b-548e-4f24-9489-7271c5d5dfee","Type":"ContainerDied","Data":"044f026b683ec5bb94ede772a08cf63539b85c76b177949c5d1caa97b9c58fc6"} Jan 26 09:17:55 crc kubenswrapper[4872]: I0126 09:17:55.006064 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="044f026b683ec5bb94ede772a08cf63539b85c76b177949c5d1caa97b9c58fc6" Jan 26 09:17:55 crc kubenswrapper[4872]: I0126 09:17:55.006098 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.123894 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2"] Jan 26 09:18:03 crc kubenswrapper[4872]: E0126 09:18:03.124926 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a3643b-548e-4f24-9489-7271c5d5dfee" containerName="util" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.124942 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a3643b-548e-4f24-9489-7271c5d5dfee" containerName="util" Jan 26 09:18:03 crc kubenswrapper[4872]: E0126 09:18:03.124954 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d625f91-d4de-4687-bdbe-8ff985d7e406" containerName="console" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.124960 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d625f91-d4de-4687-bdbe-8ff985d7e406" containerName="console" Jan 26 09:18:03 crc kubenswrapper[4872]: E0126 09:18:03.124970 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a3643b-548e-4f24-9489-7271c5d5dfee" containerName="pull" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.124976 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a3643b-548e-4f24-9489-7271c5d5dfee" containerName="pull" Jan 26 09:18:03 crc kubenswrapper[4872]: E0126 09:18:03.124987 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a3643b-548e-4f24-9489-7271c5d5dfee" containerName="extract" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.124992 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a3643b-548e-4f24-9489-7271c5d5dfee" containerName="extract" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.125103 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a3643b-548e-4f24-9489-7271c5d5dfee" containerName="extract" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.125118 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d625f91-d4de-4687-bdbe-8ff985d7e406" containerName="console" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.125587 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.130300 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.131247 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-rcbmj" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.131403 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.131545 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.146049 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2"] Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.147163 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.220064 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e-apiservice-cert\") pod \"metallb-operator-controller-manager-56ccf68bf4-prss2\" (UID: \"a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e\") " pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.220330 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e-webhook-cert\") pod \"metallb-operator-controller-manager-56ccf68bf4-prss2\" (UID: \"a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e\") " pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.220393 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g6jm\" (UniqueName: \"kubernetes.io/projected/a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e-kube-api-access-7g6jm\") pod \"metallb-operator-controller-manager-56ccf68bf4-prss2\" (UID: \"a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e\") " pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.321937 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e-webhook-cert\") pod \"metallb-operator-controller-manager-56ccf68bf4-prss2\" (UID: \"a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e\") " pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.322336 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g6jm\" (UniqueName: \"kubernetes.io/projected/a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e-kube-api-access-7g6jm\") pod \"metallb-operator-controller-manager-56ccf68bf4-prss2\" (UID: \"a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e\") " pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.322452 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e-apiservice-cert\") pod \"metallb-operator-controller-manager-56ccf68bf4-prss2\" (UID: \"a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e\") " pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.329388 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e-webhook-cert\") pod \"metallb-operator-controller-manager-56ccf68bf4-prss2\" (UID: \"a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e\") " pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.329671 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e-apiservice-cert\") pod \"metallb-operator-controller-manager-56ccf68bf4-prss2\" (UID: \"a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e\") " pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.345009 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g6jm\" (UniqueName: \"kubernetes.io/projected/a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e-kube-api-access-7g6jm\") pod \"metallb-operator-controller-manager-56ccf68bf4-prss2\" (UID: \"a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e\") " pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.403498 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8"] Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.404310 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.407041 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.407275 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.407410 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-st9dq" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.424315 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8"] Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.504583 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.524928 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8djn\" (UniqueName: \"kubernetes.io/projected/795e65d7-69c2-4449-b99d-4bbee023b065-kube-api-access-n8djn\") pod \"metallb-operator-webhook-server-6d8c8779-66lg8\" (UID: \"795e65d7-69c2-4449-b99d-4bbee023b065\") " pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.525004 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/795e65d7-69c2-4449-b99d-4bbee023b065-apiservice-cert\") pod \"metallb-operator-webhook-server-6d8c8779-66lg8\" (UID: \"795e65d7-69c2-4449-b99d-4bbee023b065\") " pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.525074 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/795e65d7-69c2-4449-b99d-4bbee023b065-webhook-cert\") pod \"metallb-operator-webhook-server-6d8c8779-66lg8\" (UID: \"795e65d7-69c2-4449-b99d-4bbee023b065\") " pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.625714 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8djn\" (UniqueName: \"kubernetes.io/projected/795e65d7-69c2-4449-b99d-4bbee023b065-kube-api-access-n8djn\") pod \"metallb-operator-webhook-server-6d8c8779-66lg8\" (UID: \"795e65d7-69c2-4449-b99d-4bbee023b065\") " pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.625772 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/795e65d7-69c2-4449-b99d-4bbee023b065-apiservice-cert\") pod \"metallb-operator-webhook-server-6d8c8779-66lg8\" (UID: \"795e65d7-69c2-4449-b99d-4bbee023b065\") " pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.625840 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/795e65d7-69c2-4449-b99d-4bbee023b065-webhook-cert\") pod \"metallb-operator-webhook-server-6d8c8779-66lg8\" (UID: \"795e65d7-69c2-4449-b99d-4bbee023b065\") " pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.633006 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/795e65d7-69c2-4449-b99d-4bbee023b065-apiservice-cert\") pod \"metallb-operator-webhook-server-6d8c8779-66lg8\" (UID: \"795e65d7-69c2-4449-b99d-4bbee023b065\") " pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.633080 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/795e65d7-69c2-4449-b99d-4bbee023b065-webhook-cert\") pod \"metallb-operator-webhook-server-6d8c8779-66lg8\" (UID: \"795e65d7-69c2-4449-b99d-4bbee023b065\") " pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.648261 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8djn\" (UniqueName: \"kubernetes.io/projected/795e65d7-69c2-4449-b99d-4bbee023b065-kube-api-access-n8djn\") pod \"metallb-operator-webhook-server-6d8c8779-66lg8\" (UID: \"795e65d7-69c2-4449-b99d-4bbee023b065\") " pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:03 crc kubenswrapper[4872]: I0126 09:18:03.724611 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:04 crc kubenswrapper[4872]: I0126 09:18:04.009478 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2"] Jan 26 09:18:04 crc kubenswrapper[4872]: I0126 09:18:04.087182 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" event={"ID":"a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e","Type":"ContainerStarted","Data":"6cffae8e432052295a50043b950d3a2f8775ef27380bc9b9df48d675cfd32b37"} Jan 26 09:18:04 crc kubenswrapper[4872]: I0126 09:18:04.171559 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8"] Jan 26 09:18:04 crc kubenswrapper[4872]: W0126 09:18:04.174387 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod795e65d7_69c2_4449_b99d_4bbee023b065.slice/crio-cce202eb9f98b6b78370c7feb040dd6eee6faf93e98e113f3cbba0fee4282bb1 WatchSource:0}: Error finding container cce202eb9f98b6b78370c7feb040dd6eee6faf93e98e113f3cbba0fee4282bb1: Status 404 returned error can't find the container with id cce202eb9f98b6b78370c7feb040dd6eee6faf93e98e113f3cbba0fee4282bb1 Jan 26 09:18:05 crc kubenswrapper[4872]: I0126 09:18:05.094533 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" event={"ID":"795e65d7-69c2-4449-b99d-4bbee023b065","Type":"ContainerStarted","Data":"cce202eb9f98b6b78370c7feb040dd6eee6faf93e98e113f3cbba0fee4282bb1"} Jan 26 09:18:11 crc kubenswrapper[4872]: I0126 09:18:11.194151 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" event={"ID":"a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e","Type":"ContainerStarted","Data":"e4c8ad029e9c0eca778eafe72a71a94616ef81596475f8558b7a798bbf4fc8b6"} Jan 26 09:18:11 crc kubenswrapper[4872]: I0126 09:18:11.195919 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:11 crc kubenswrapper[4872]: I0126 09:18:11.198083 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" event={"ID":"795e65d7-69c2-4449-b99d-4bbee023b065","Type":"ContainerStarted","Data":"1b00e7d7ca4e1131dce6ca4a26b2fe7612251bbb6bc95f9332d42db5aa02ba1b"} Jan 26 09:18:11 crc kubenswrapper[4872]: I0126 09:18:11.198362 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:11 crc kubenswrapper[4872]: I0126 09:18:11.246620 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" podStartSLOduration=1.894864374 podStartE2EDuration="8.246592641s" podCreationTimestamp="2026-01-26 09:18:03 +0000 UTC" firstStartedPulling="2026-01-26 09:18:04.027269249 +0000 UTC m=+617.336109060" lastFinishedPulling="2026-01-26 09:18:10.378997536 +0000 UTC m=+623.687837327" observedRunningTime="2026-01-26 09:18:11.221261168 +0000 UTC m=+624.530100969" watchObservedRunningTime="2026-01-26 09:18:11.246592641 +0000 UTC m=+624.555432452" Jan 26 09:18:11 crc kubenswrapper[4872]: I0126 09:18:11.246770 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" podStartSLOduration=2.022275729 podStartE2EDuration="8.246762495s" podCreationTimestamp="2026-01-26 09:18:03 +0000 UTC" firstStartedPulling="2026-01-26 09:18:04.178096397 +0000 UTC m=+617.486936198" lastFinishedPulling="2026-01-26 09:18:10.402583133 +0000 UTC m=+623.711422964" observedRunningTime="2026-01-26 09:18:11.242949484 +0000 UTC m=+624.551789295" watchObservedRunningTime="2026-01-26 09:18:11.246762495 +0000 UTC m=+624.555602296" Jan 26 09:18:23 crc kubenswrapper[4872]: I0126 09:18:23.740893 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6d8c8779-66lg8" Jan 26 09:18:43 crc kubenswrapper[4872]: I0126 09:18:43.508539 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-56ccf68bf4-prss2" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.233993 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-zmwfb"] Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.237769 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.239515 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv"] Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.240616 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-ggbl7" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.241029 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.241307 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.241537 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.242452 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.245090 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv"] Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.341497 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-c2rkk"] Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.342621 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-c2rkk" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.345306 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.345520 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.345645 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.345763 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-zv2px" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.379180 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x892g\" (UniqueName: \"kubernetes.io/projected/965facdf-05ce-4a41-8a03-cf20d04932d3-kube-api-access-x892g\") pod \"frr-k8s-webhook-server-7df86c4f6c-m7pqv\" (UID: \"965facdf-05ce-4a41-8a03-cf20d04932d3\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.379246 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/965facdf-05ce-4a41-8a03-cf20d04932d3-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-m7pqv\" (UID: \"965facdf-05ce-4a41-8a03-cf20d04932d3\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.379288 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/bc48ed53-17c1-43f3-9b31-d04c7d510282-frr-conf\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.379312 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/bc48ed53-17c1-43f3-9b31-d04c7d510282-reloader\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.379331 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/bc48ed53-17c1-43f3-9b31-d04c7d510282-metrics\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.379349 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/bc48ed53-17c1-43f3-9b31-d04c7d510282-frr-sockets\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.379376 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vszsm\" (UniqueName: \"kubernetes.io/projected/bc48ed53-17c1-43f3-9b31-d04c7d510282-kube-api-access-vszsm\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.379397 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48ed53-17c1-43f3-9b31-d04c7d510282-metrics-certs\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.379411 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/bc48ed53-17c1-43f3-9b31-d04c7d510282-frr-startup\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.397374 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6968d8fdc4-whrlc"] Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.398676 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.412849 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-whrlc"] Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.421339 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.488921 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-metrics-certs\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489003 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/bc48ed53-17c1-43f3-9b31-d04c7d510282-reloader\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489040 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/bc48ed53-17c1-43f3-9b31-d04c7d510282-metrics\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489063 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1326511a-69a9-48da-850a-c05a3fa98fad-metrics-certs\") pod \"controller-6968d8fdc4-whrlc\" (UID: \"1326511a-69a9-48da-850a-c05a3fa98fad\") " pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489080 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh478\" (UniqueName: \"kubernetes.io/projected/1326511a-69a9-48da-850a-c05a3fa98fad-kube-api-access-lh478\") pod \"controller-6968d8fdc4-whrlc\" (UID: \"1326511a-69a9-48da-850a-c05a3fa98fad\") " pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489103 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/bc48ed53-17c1-43f3-9b31-d04c7d510282-frr-sockets\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489125 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt76j\" (UniqueName: \"kubernetes.io/projected/816522a7-9120-431e-9f3d-2c66f61d050c-kube-api-access-lt76j\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489144 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-memberlist\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489161 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vszsm\" (UniqueName: \"kubernetes.io/projected/bc48ed53-17c1-43f3-9b31-d04c7d510282-kube-api-access-vszsm\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489180 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/816522a7-9120-431e-9f3d-2c66f61d050c-metallb-excludel2\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489202 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48ed53-17c1-43f3-9b31-d04c7d510282-metrics-certs\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489220 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/bc48ed53-17c1-43f3-9b31-d04c7d510282-frr-startup\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489248 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x892g\" (UniqueName: \"kubernetes.io/projected/965facdf-05ce-4a41-8a03-cf20d04932d3-kube-api-access-x892g\") pod \"frr-k8s-webhook-server-7df86c4f6c-m7pqv\" (UID: \"965facdf-05ce-4a41-8a03-cf20d04932d3\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489272 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/965facdf-05ce-4a41-8a03-cf20d04932d3-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-m7pqv\" (UID: \"965facdf-05ce-4a41-8a03-cf20d04932d3\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489307 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1326511a-69a9-48da-850a-c05a3fa98fad-cert\") pod \"controller-6968d8fdc4-whrlc\" (UID: \"1326511a-69a9-48da-850a-c05a3fa98fad\") " pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489329 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/bc48ed53-17c1-43f3-9b31-d04c7d510282-frr-conf\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.489988 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/bc48ed53-17c1-43f3-9b31-d04c7d510282-frr-conf\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.490241 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/bc48ed53-17c1-43f3-9b31-d04c7d510282-reloader\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.492759 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/bc48ed53-17c1-43f3-9b31-d04c7d510282-frr-startup\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: E0126 09:18:44.492922 4872 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Jan 26 09:18:44 crc kubenswrapper[4872]: E0126 09:18:44.492988 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bc48ed53-17c1-43f3-9b31-d04c7d510282-metrics-certs podName:bc48ed53-17c1-43f3-9b31-d04c7d510282 nodeName:}" failed. No retries permitted until 2026-01-26 09:18:44.992965067 +0000 UTC m=+658.301804868 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/bc48ed53-17c1-43f3-9b31-d04c7d510282-metrics-certs") pod "frr-k8s-zmwfb" (UID: "bc48ed53-17c1-43f3-9b31-d04c7d510282") : secret "frr-k8s-certs-secret" not found Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.493433 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/bc48ed53-17c1-43f3-9b31-d04c7d510282-frr-sockets\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.531325 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/965facdf-05ce-4a41-8a03-cf20d04932d3-cert\") pod \"frr-k8s-webhook-server-7df86c4f6c-m7pqv\" (UID: \"965facdf-05ce-4a41-8a03-cf20d04932d3\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.540536 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vszsm\" (UniqueName: \"kubernetes.io/projected/bc48ed53-17c1-43f3-9b31-d04c7d510282-kube-api-access-vszsm\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.542073 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x892g\" (UniqueName: \"kubernetes.io/projected/965facdf-05ce-4a41-8a03-cf20d04932d3-kube-api-access-x892g\") pod \"frr-k8s-webhook-server-7df86c4f6c-m7pqv\" (UID: \"965facdf-05ce-4a41-8a03-cf20d04932d3\") " pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.570269 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.590114 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1326511a-69a9-48da-850a-c05a3fa98fad-metrics-certs\") pod \"controller-6968d8fdc4-whrlc\" (UID: \"1326511a-69a9-48da-850a-c05a3fa98fad\") " pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.590171 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh478\" (UniqueName: \"kubernetes.io/projected/1326511a-69a9-48da-850a-c05a3fa98fad-kube-api-access-lh478\") pod \"controller-6968d8fdc4-whrlc\" (UID: \"1326511a-69a9-48da-850a-c05a3fa98fad\") " pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.590196 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt76j\" (UniqueName: \"kubernetes.io/projected/816522a7-9120-431e-9f3d-2c66f61d050c-kube-api-access-lt76j\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.590225 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-memberlist\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.590258 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/816522a7-9120-431e-9f3d-2c66f61d050c-metallb-excludel2\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.590326 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1326511a-69a9-48da-850a-c05a3fa98fad-cert\") pod \"controller-6968d8fdc4-whrlc\" (UID: \"1326511a-69a9-48da-850a-c05a3fa98fad\") " pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.590355 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-metrics-certs\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:44 crc kubenswrapper[4872]: E0126 09:18:44.590686 4872 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Jan 26 09:18:44 crc kubenswrapper[4872]: E0126 09:18:44.590756 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-metrics-certs podName:816522a7-9120-431e-9f3d-2c66f61d050c nodeName:}" failed. No retries permitted until 2026-01-26 09:18:45.090734321 +0000 UTC m=+658.399574122 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-metrics-certs") pod "speaker-c2rkk" (UID: "816522a7-9120-431e-9f3d-2c66f61d050c") : secret "speaker-certs-secret" not found Jan 26 09:18:44 crc kubenswrapper[4872]: E0126 09:18:44.591037 4872 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Jan 26 09:18:44 crc kubenswrapper[4872]: E0126 09:18:44.591065 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1326511a-69a9-48da-850a-c05a3fa98fad-metrics-certs podName:1326511a-69a9-48da-850a-c05a3fa98fad nodeName:}" failed. No retries permitted until 2026-01-26 09:18:45.091057759 +0000 UTC m=+658.399897560 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/1326511a-69a9-48da-850a-c05a3fa98fad-metrics-certs") pod "controller-6968d8fdc4-whrlc" (UID: "1326511a-69a9-48da-850a-c05a3fa98fad") : secret "controller-certs-secret" not found Jan 26 09:18:44 crc kubenswrapper[4872]: E0126 09:18:44.591364 4872 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 26 09:18:44 crc kubenswrapper[4872]: E0126 09:18:44.591401 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-memberlist podName:816522a7-9120-431e-9f3d-2c66f61d050c nodeName:}" failed. No retries permitted until 2026-01-26 09:18:45.091389347 +0000 UTC m=+658.400229148 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-memberlist") pod "speaker-c2rkk" (UID: "816522a7-9120-431e-9f3d-2c66f61d050c") : secret "metallb-memberlist" not found Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.592089 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/816522a7-9120-431e-9f3d-2c66f61d050c-metallb-excludel2\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.615819 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh478\" (UniqueName: \"kubernetes.io/projected/1326511a-69a9-48da-850a-c05a3fa98fad-kube-api-access-lh478\") pod \"controller-6968d8fdc4-whrlc\" (UID: \"1326511a-69a9-48da-850a-c05a3fa98fad\") " pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.616402 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1326511a-69a9-48da-850a-c05a3fa98fad-cert\") pod \"controller-6968d8fdc4-whrlc\" (UID: \"1326511a-69a9-48da-850a-c05a3fa98fad\") " pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.620275 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt76j\" (UniqueName: \"kubernetes.io/projected/816522a7-9120-431e-9f3d-2c66f61d050c-kube-api-access-lt76j\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.880334 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/bc48ed53-17c1-43f3-9b31-d04c7d510282-metrics\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.930305 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv"] Jan 26 09:18:44 crc kubenswrapper[4872]: I0126 09:18:44.995911 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48ed53-17c1-43f3-9b31-d04c7d510282-metrics-certs\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:45 crc kubenswrapper[4872]: I0126 09:18:45.009882 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/bc48ed53-17c1-43f3-9b31-d04c7d510282-metrics-certs\") pod \"frr-k8s-zmwfb\" (UID: \"bc48ed53-17c1-43f3-9b31-d04c7d510282\") " pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:45 crc kubenswrapper[4872]: I0126 09:18:45.097859 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-metrics-certs\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:45 crc kubenswrapper[4872]: I0126 09:18:45.097936 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1326511a-69a9-48da-850a-c05a3fa98fad-metrics-certs\") pod \"controller-6968d8fdc4-whrlc\" (UID: \"1326511a-69a9-48da-850a-c05a3fa98fad\") " pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:45 crc kubenswrapper[4872]: I0126 09:18:45.097980 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-memberlist\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:45 crc kubenswrapper[4872]: E0126 09:18:45.098213 4872 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Jan 26 09:18:45 crc kubenswrapper[4872]: E0126 09:18:45.098297 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-memberlist podName:816522a7-9120-431e-9f3d-2c66f61d050c nodeName:}" failed. No retries permitted until 2026-01-26 09:18:46.098271899 +0000 UTC m=+659.407111700 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-memberlist") pod "speaker-c2rkk" (UID: "816522a7-9120-431e-9f3d-2c66f61d050c") : secret "metallb-memberlist" not found Jan 26 09:18:45 crc kubenswrapper[4872]: I0126 09:18:45.102230 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1326511a-69a9-48da-850a-c05a3fa98fad-metrics-certs\") pod \"controller-6968d8fdc4-whrlc\" (UID: \"1326511a-69a9-48da-850a-c05a3fa98fad\") " pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:45 crc kubenswrapper[4872]: I0126 09:18:45.103423 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-metrics-certs\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:45 crc kubenswrapper[4872]: I0126 09:18:45.159822 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:45 crc kubenswrapper[4872]: I0126 09:18:45.324021 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:45 crc kubenswrapper[4872]: I0126 09:18:45.435330 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmwfb" event={"ID":"bc48ed53-17c1-43f3-9b31-d04c7d510282","Type":"ContainerStarted","Data":"b4a4b0e7d359bf9d2ab603217efa41fe1729f8b7f52ffbad1f0d39cdbd45545d"} Jan 26 09:18:45 crc kubenswrapper[4872]: I0126 09:18:45.442521 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" event={"ID":"965facdf-05ce-4a41-8a03-cf20d04932d3","Type":"ContainerStarted","Data":"fff079c7e6107ec6b50b58e9b753f88bac8692181990ea026bf2da3ec1d2e8dc"} Jan 26 09:18:45 crc kubenswrapper[4872]: I0126 09:18:45.542361 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6968d8fdc4-whrlc"] Jan 26 09:18:45 crc kubenswrapper[4872]: W0126 09:18:45.566198 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1326511a_69a9_48da_850a_c05a3fa98fad.slice/crio-ae4be710349030afecadc086c4ea71264bbe99c8f09eca0e5d2dbb3990d54d02 WatchSource:0}: Error finding container ae4be710349030afecadc086c4ea71264bbe99c8f09eca0e5d2dbb3990d54d02: Status 404 returned error can't find the container with id ae4be710349030afecadc086c4ea71264bbe99c8f09eca0e5d2dbb3990d54d02 Jan 26 09:18:46 crc kubenswrapper[4872]: I0126 09:18:46.112140 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-memberlist\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:46 crc kubenswrapper[4872]: I0126 09:18:46.120417 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/816522a7-9120-431e-9f3d-2c66f61d050c-memberlist\") pod \"speaker-c2rkk\" (UID: \"816522a7-9120-431e-9f3d-2c66f61d050c\") " pod="metallb-system/speaker-c2rkk" Jan 26 09:18:46 crc kubenswrapper[4872]: I0126 09:18:46.294433 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-c2rkk" Jan 26 09:18:46 crc kubenswrapper[4872]: I0126 09:18:46.458163 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-whrlc" event={"ID":"1326511a-69a9-48da-850a-c05a3fa98fad","Type":"ContainerStarted","Data":"0890193dffc979dd01a38fec7e6a4c7f0f0cd3a7414e6e565c7b00fee3090560"} Jan 26 09:18:46 crc kubenswrapper[4872]: I0126 09:18:46.458422 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-whrlc" event={"ID":"1326511a-69a9-48da-850a-c05a3fa98fad","Type":"ContainerStarted","Data":"ca4a6b29e77bf11e0332c3e202913aef4c0688ba2c2fc3388aeb9e1675df6b98"} Jan 26 09:18:46 crc kubenswrapper[4872]: I0126 09:18:46.458446 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6968d8fdc4-whrlc" event={"ID":"1326511a-69a9-48da-850a-c05a3fa98fad","Type":"ContainerStarted","Data":"ae4be710349030afecadc086c4ea71264bbe99c8f09eca0e5d2dbb3990d54d02"} Jan 26 09:18:46 crc kubenswrapper[4872]: I0126 09:18:46.458489 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:46 crc kubenswrapper[4872]: I0126 09:18:46.463032 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-c2rkk" event={"ID":"816522a7-9120-431e-9f3d-2c66f61d050c","Type":"ContainerStarted","Data":"76055cf9febd434fb29128751d62e0737ea5a196a331c9d56ce2e66acfa39f76"} Jan 26 09:18:46 crc kubenswrapper[4872]: I0126 09:18:46.483227 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6968d8fdc4-whrlc" podStartSLOduration=2.483196506 podStartE2EDuration="2.483196506s" podCreationTimestamp="2026-01-26 09:18:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:18:46.48065207 +0000 UTC m=+659.789491871" watchObservedRunningTime="2026-01-26 09:18:46.483196506 +0000 UTC m=+659.792036307" Jan 26 09:18:47 crc kubenswrapper[4872]: I0126 09:18:47.494390 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-c2rkk" event={"ID":"816522a7-9120-431e-9f3d-2c66f61d050c","Type":"ContainerStarted","Data":"d0ae368287f078a3afa989310c2c7c8dae64708a8f66c416c6d97ad742461ef1"} Jan 26 09:18:47 crc kubenswrapper[4872]: I0126 09:18:47.495353 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-c2rkk" event={"ID":"816522a7-9120-431e-9f3d-2c66f61d050c","Type":"ContainerStarted","Data":"5f8f2131aedb85325f989c5b1ada978a77a534535881989a37715246b2b3b6c0"} Jan 26 09:18:47 crc kubenswrapper[4872]: I0126 09:18:47.495384 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-c2rkk" Jan 26 09:18:47 crc kubenswrapper[4872]: I0126 09:18:47.514873 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-c2rkk" podStartSLOduration=3.514852739 podStartE2EDuration="3.514852739s" podCreationTimestamp="2026-01-26 09:18:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:18:47.513840802 +0000 UTC m=+660.822680613" watchObservedRunningTime="2026-01-26 09:18:47.514852739 +0000 UTC m=+660.823692540" Jan 26 09:18:54 crc kubenswrapper[4872]: I0126 09:18:54.584470 4872 generic.go:334] "Generic (PLEG): container finished" podID="bc48ed53-17c1-43f3-9b31-d04c7d510282" containerID="db62f35da010eab9c353f5fd34ff0488dd3b5839f3e15dcd36748c42f1e098ae" exitCode=0 Jan 26 09:18:54 crc kubenswrapper[4872]: I0126 09:18:54.584634 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmwfb" event={"ID":"bc48ed53-17c1-43f3-9b31-d04c7d510282","Type":"ContainerDied","Data":"db62f35da010eab9c353f5fd34ff0488dd3b5839f3e15dcd36748c42f1e098ae"} Jan 26 09:18:54 crc kubenswrapper[4872]: I0126 09:18:54.590109 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" event={"ID":"965facdf-05ce-4a41-8a03-cf20d04932d3","Type":"ContainerStarted","Data":"730eb9c2045392060c7dc776b02a4af024dda9c306dd6395819b629150ad01aa"} Jan 26 09:18:54 crc kubenswrapper[4872]: I0126 09:18:54.590401 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" Jan 26 09:18:54 crc kubenswrapper[4872]: I0126 09:18:54.648451 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" podStartSLOduration=1.6854388070000001 podStartE2EDuration="10.648427871s" podCreationTimestamp="2026-01-26 09:18:44 +0000 UTC" firstStartedPulling="2026-01-26 09:18:44.936374534 +0000 UTC m=+658.245214335" lastFinishedPulling="2026-01-26 09:18:53.899363598 +0000 UTC m=+667.208203399" observedRunningTime="2026-01-26 09:18:54.647114026 +0000 UTC m=+667.955953837" watchObservedRunningTime="2026-01-26 09:18:54.648427871 +0000 UTC m=+667.957267682" Jan 26 09:18:55 crc kubenswrapper[4872]: I0126 09:18:55.329703 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6968d8fdc4-whrlc" Jan 26 09:18:55 crc kubenswrapper[4872]: I0126 09:18:55.602687 4872 generic.go:334] "Generic (PLEG): container finished" podID="bc48ed53-17c1-43f3-9b31-d04c7d510282" containerID="19866e02b0c862494b022a4034c5a2d3600749e5e840bdce90daf4f09fd59e3b" exitCode=0 Jan 26 09:18:55 crc kubenswrapper[4872]: I0126 09:18:55.602774 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmwfb" event={"ID":"bc48ed53-17c1-43f3-9b31-d04c7d510282","Type":"ContainerDied","Data":"19866e02b0c862494b022a4034c5a2d3600749e5e840bdce90daf4f09fd59e3b"} Jan 26 09:18:56 crc kubenswrapper[4872]: I0126 09:18:56.301311 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-c2rkk" Jan 26 09:18:56 crc kubenswrapper[4872]: I0126 09:18:56.613982 4872 generic.go:334] "Generic (PLEG): container finished" podID="bc48ed53-17c1-43f3-9b31-d04c7d510282" containerID="6a94e04db68f313a8bca9684c55c05553ec2b826cd6fd031446d3db01602104d" exitCode=0 Jan 26 09:18:56 crc kubenswrapper[4872]: I0126 09:18:56.614051 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmwfb" event={"ID":"bc48ed53-17c1-43f3-9b31-d04c7d510282","Type":"ContainerDied","Data":"6a94e04db68f313a8bca9684c55c05553ec2b826cd6fd031446d3db01602104d"} Jan 26 09:18:57 crc kubenswrapper[4872]: I0126 09:18:57.631486 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmwfb" event={"ID":"bc48ed53-17c1-43f3-9b31-d04c7d510282","Type":"ContainerStarted","Data":"e198c1cf9aabd10e384f592931eeb4fb53b2e981581017e97b2cb17ef96c321e"} Jan 26 09:18:57 crc kubenswrapper[4872]: I0126 09:18:57.632869 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmwfb" event={"ID":"bc48ed53-17c1-43f3-9b31-d04c7d510282","Type":"ContainerStarted","Data":"0f0b457118b2b62e2481a674d129caa30101abc09838e0e55832c061b13f6e6c"} Jan 26 09:18:57 crc kubenswrapper[4872]: I0126 09:18:57.632915 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmwfb" event={"ID":"bc48ed53-17c1-43f3-9b31-d04c7d510282","Type":"ContainerStarted","Data":"d34f69792b2c23589153f2683df58e2166978693b6d8b0445928f16236f34c2e"} Jan 26 09:18:57 crc kubenswrapper[4872]: I0126 09:18:57.632931 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmwfb" event={"ID":"bc48ed53-17c1-43f3-9b31-d04c7d510282","Type":"ContainerStarted","Data":"4ca4f56ce2d9d2858c763dad1ec53ebc8a06d9e0d4766d61bbec1bd198bd5802"} Jan 26 09:18:57 crc kubenswrapper[4872]: I0126 09:18:57.632942 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmwfb" event={"ID":"bc48ed53-17c1-43f3-9b31-d04c7d510282","Type":"ContainerStarted","Data":"db8a15528b132ccb1c44399b355c6c08ec6f4cc9886cea39151ec1c596e20812"} Jan 26 09:18:58 crc kubenswrapper[4872]: I0126 09:18:58.647212 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-zmwfb" event={"ID":"bc48ed53-17c1-43f3-9b31-d04c7d510282","Type":"ContainerStarted","Data":"faa08bea75b93019aa1eeabcb9db3d9252a7ca51cd2358532bebb9d9ac1e8f34"} Jan 26 09:18:58 crc kubenswrapper[4872]: I0126 09:18:58.647450 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:18:58 crc kubenswrapper[4872]: I0126 09:18:58.697096 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-zmwfb" podStartSLOduration=6.087672597 podStartE2EDuration="14.697073758s" podCreationTimestamp="2026-01-26 09:18:44 +0000 UTC" firstStartedPulling="2026-01-26 09:18:45.273764102 +0000 UTC m=+658.582603903" lastFinishedPulling="2026-01-26 09:18:53.883165263 +0000 UTC m=+667.192005064" observedRunningTime="2026-01-26 09:18:58.694391347 +0000 UTC m=+672.003231158" watchObservedRunningTime="2026-01-26 09:18:58.697073758 +0000 UTC m=+672.005913559" Jan 26 09:18:59 crc kubenswrapper[4872]: I0126 09:18:59.070652 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-t78xn"] Jan 26 09:18:59 crc kubenswrapper[4872]: I0126 09:18:59.072143 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t78xn" Jan 26 09:18:59 crc kubenswrapper[4872]: I0126 09:18:59.075250 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Jan 26 09:18:59 crc kubenswrapper[4872]: I0126 09:18:59.075252 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Jan 26 09:18:59 crc kubenswrapper[4872]: I0126 09:18:59.076279 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-bzxph" Jan 26 09:18:59 crc kubenswrapper[4872]: I0126 09:18:59.094296 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t78xn"] Jan 26 09:18:59 crc kubenswrapper[4872]: I0126 09:18:59.150528 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmkpd\" (UniqueName: \"kubernetes.io/projected/4268cbf4-e6ce-41f3-8291-728b5516ad10-kube-api-access-nmkpd\") pod \"openstack-operator-index-t78xn\" (UID: \"4268cbf4-e6ce-41f3-8291-728b5516ad10\") " pod="openstack-operators/openstack-operator-index-t78xn" Jan 26 09:18:59 crc kubenswrapper[4872]: I0126 09:18:59.252153 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmkpd\" (UniqueName: \"kubernetes.io/projected/4268cbf4-e6ce-41f3-8291-728b5516ad10-kube-api-access-nmkpd\") pod \"openstack-operator-index-t78xn\" (UID: \"4268cbf4-e6ce-41f3-8291-728b5516ad10\") " pod="openstack-operators/openstack-operator-index-t78xn" Jan 26 09:18:59 crc kubenswrapper[4872]: I0126 09:18:59.272162 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmkpd\" (UniqueName: \"kubernetes.io/projected/4268cbf4-e6ce-41f3-8291-728b5516ad10-kube-api-access-nmkpd\") pod \"openstack-operator-index-t78xn\" (UID: \"4268cbf4-e6ce-41f3-8291-728b5516ad10\") " pod="openstack-operators/openstack-operator-index-t78xn" Jan 26 09:18:59 crc kubenswrapper[4872]: I0126 09:18:59.415399 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t78xn" Jan 26 09:19:00 crc kubenswrapper[4872]: I0126 09:19:00.152377 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t78xn"] Jan 26 09:19:00 crc kubenswrapper[4872]: I0126 09:19:00.163896 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:19:00 crc kubenswrapper[4872]: W0126 09:19:00.167112 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4268cbf4_e6ce_41f3_8291_728b5516ad10.slice/crio-e148e38675b6d6ec3eeb6cddbe33d8046c9189d6d41525390240e06fe0f9e5db WatchSource:0}: Error finding container e148e38675b6d6ec3eeb6cddbe33d8046c9189d6d41525390240e06fe0f9e5db: Status 404 returned error can't find the container with id e148e38675b6d6ec3eeb6cddbe33d8046c9189d6d41525390240e06fe0f9e5db Jan 26 09:19:00 crc kubenswrapper[4872]: I0126 09:19:00.217242 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:19:00 crc kubenswrapper[4872]: I0126 09:19:00.664107 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t78xn" event={"ID":"4268cbf4-e6ce-41f3-8291-728b5516ad10","Type":"ContainerStarted","Data":"e148e38675b6d6ec3eeb6cddbe33d8046c9189d6d41525390240e06fe0f9e5db"} Jan 26 09:19:02 crc kubenswrapper[4872]: I0126 09:19:02.439191 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-t78xn"] Jan 26 09:19:03 crc kubenswrapper[4872]: I0126 09:19:03.048878 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-psjm6"] Jan 26 09:19:03 crc kubenswrapper[4872]: I0126 09:19:03.050089 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-psjm6" Jan 26 09:19:03 crc kubenswrapper[4872]: I0126 09:19:03.055649 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-psjm6"] Jan 26 09:19:03 crc kubenswrapper[4872]: I0126 09:19:03.106644 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gr7z\" (UniqueName: \"kubernetes.io/projected/1b7d4719-ae54-402a-83c2-39ca5edb30d0-kube-api-access-9gr7z\") pod \"openstack-operator-index-psjm6\" (UID: \"1b7d4719-ae54-402a-83c2-39ca5edb30d0\") " pod="openstack-operators/openstack-operator-index-psjm6" Jan 26 09:19:03 crc kubenswrapper[4872]: I0126 09:19:03.209446 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gr7z\" (UniqueName: \"kubernetes.io/projected/1b7d4719-ae54-402a-83c2-39ca5edb30d0-kube-api-access-9gr7z\") pod \"openstack-operator-index-psjm6\" (UID: \"1b7d4719-ae54-402a-83c2-39ca5edb30d0\") " pod="openstack-operators/openstack-operator-index-psjm6" Jan 26 09:19:03 crc kubenswrapper[4872]: I0126 09:19:03.235583 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gr7z\" (UniqueName: \"kubernetes.io/projected/1b7d4719-ae54-402a-83c2-39ca5edb30d0-kube-api-access-9gr7z\") pod \"openstack-operator-index-psjm6\" (UID: \"1b7d4719-ae54-402a-83c2-39ca5edb30d0\") " pod="openstack-operators/openstack-operator-index-psjm6" Jan 26 09:19:03 crc kubenswrapper[4872]: I0126 09:19:03.375554 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-psjm6" Jan 26 09:19:03 crc kubenswrapper[4872]: I0126 09:19:03.618194 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-psjm6"] Jan 26 09:19:03 crc kubenswrapper[4872]: I0126 09:19:03.687770 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-psjm6" event={"ID":"1b7d4719-ae54-402a-83c2-39ca5edb30d0","Type":"ContainerStarted","Data":"4a682393b9fa066393f3b6ef34a45d195b11d4ad9a89d586d1a57938dbb39929"} Jan 26 09:19:03 crc kubenswrapper[4872]: I0126 09:19:03.691373 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t78xn" event={"ID":"4268cbf4-e6ce-41f3-8291-728b5516ad10","Type":"ContainerStarted","Data":"dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5"} Jan 26 09:19:03 crc kubenswrapper[4872]: I0126 09:19:03.691656 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-t78xn" podUID="4268cbf4-e6ce-41f3-8291-728b5516ad10" containerName="registry-server" containerID="cri-o://dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5" gracePeriod=2 Jan 26 09:19:03 crc kubenswrapper[4872]: I0126 09:19:03.716611 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-t78xn" podStartSLOduration=1.5635179209999999 podStartE2EDuration="4.716585483s" podCreationTimestamp="2026-01-26 09:18:59 +0000 UTC" firstStartedPulling="2026-01-26 09:19:00.169000405 +0000 UTC m=+673.477840206" lastFinishedPulling="2026-01-26 09:19:03.322067967 +0000 UTC m=+676.630907768" observedRunningTime="2026-01-26 09:19:03.713920103 +0000 UTC m=+677.022759914" watchObservedRunningTime="2026-01-26 09:19:03.716585483 +0000 UTC m=+677.025425284" Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.007339 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t78xn" Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.031046 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmkpd\" (UniqueName: \"kubernetes.io/projected/4268cbf4-e6ce-41f3-8291-728b5516ad10-kube-api-access-nmkpd\") pod \"4268cbf4-e6ce-41f3-8291-728b5516ad10\" (UID: \"4268cbf4-e6ce-41f3-8291-728b5516ad10\") " Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.038250 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4268cbf4-e6ce-41f3-8291-728b5516ad10-kube-api-access-nmkpd" (OuterVolumeSpecName: "kube-api-access-nmkpd") pod "4268cbf4-e6ce-41f3-8291-728b5516ad10" (UID: "4268cbf4-e6ce-41f3-8291-728b5516ad10"). InnerVolumeSpecName "kube-api-access-nmkpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.134958 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmkpd\" (UniqueName: \"kubernetes.io/projected/4268cbf4-e6ce-41f3-8291-728b5516ad10-kube-api-access-nmkpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.580596 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7df86c4f6c-m7pqv" Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.702452 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-psjm6" event={"ID":"1b7d4719-ae54-402a-83c2-39ca5edb30d0","Type":"ContainerStarted","Data":"450ebfd37ed5f6178e86a817628df6a05c2a1adfdb7367511c1afb5267e36125"} Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.706423 4872 generic.go:334] "Generic (PLEG): container finished" podID="4268cbf4-e6ce-41f3-8291-728b5516ad10" containerID="dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5" exitCode=0 Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.706494 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t78xn" event={"ID":"4268cbf4-e6ce-41f3-8291-728b5516ad10","Type":"ContainerDied","Data":"dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5"} Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.706529 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t78xn" event={"ID":"4268cbf4-e6ce-41f3-8291-728b5516ad10","Type":"ContainerDied","Data":"e148e38675b6d6ec3eeb6cddbe33d8046c9189d6d41525390240e06fe0f9e5db"} Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.706554 4872 scope.go:117] "RemoveContainer" containerID="dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5" Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.706692 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t78xn" Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.728505 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-psjm6" podStartSLOduration=1.653607813 podStartE2EDuration="1.728475177s" podCreationTimestamp="2026-01-26 09:19:03 +0000 UTC" firstStartedPulling="2026-01-26 09:19:03.629524109 +0000 UTC m=+676.938363910" lastFinishedPulling="2026-01-26 09:19:03.704391463 +0000 UTC m=+677.013231274" observedRunningTime="2026-01-26 09:19:04.72208748 +0000 UTC m=+678.030927311" watchObservedRunningTime="2026-01-26 09:19:04.728475177 +0000 UTC m=+678.037314998" Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.746460 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-t78xn"] Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.749573 4872 scope.go:117] "RemoveContainer" containerID="dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5" Jan 26 09:19:04 crc kubenswrapper[4872]: E0126 09:19:04.750176 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5\": container with ID starting with dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5 not found: ID does not exist" containerID="dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5" Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.750218 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5"} err="failed to get container status \"dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5\": rpc error: code = NotFound desc = could not find container \"dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5\": container with ID starting with dbe393b7012e26ce17bd119f56c5e066c5cef5e1750b279114f91d434f35d1b5 not found: ID does not exist" Jan 26 09:19:04 crc kubenswrapper[4872]: I0126 09:19:04.751196 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-t78xn"] Jan 26 09:19:05 crc kubenswrapper[4872]: I0126 09:19:05.198698 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4268cbf4-e6ce-41f3-8291-728b5516ad10" path="/var/lib/kubelet/pods/4268cbf4-e6ce-41f3-8291-728b5516ad10/volumes" Jan 26 09:19:13 crc kubenswrapper[4872]: I0126 09:19:13.376049 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-psjm6" Jan 26 09:19:13 crc kubenswrapper[4872]: I0126 09:19:13.377056 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-psjm6" Jan 26 09:19:13 crc kubenswrapper[4872]: I0126 09:19:13.413333 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-psjm6" Jan 26 09:19:13 crc kubenswrapper[4872]: I0126 09:19:13.814766 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-psjm6" Jan 26 09:19:15 crc kubenswrapper[4872]: I0126 09:19:15.164479 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-zmwfb" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.094422 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz"] Jan 26 09:19:21 crc kubenswrapper[4872]: E0126 09:19:21.095238 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4268cbf4-e6ce-41f3-8291-728b5516ad10" containerName="registry-server" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.095252 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="4268cbf4-e6ce-41f3-8291-728b5516ad10" containerName="registry-server" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.095387 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="4268cbf4-e6ce-41f3-8291-728b5516ad10" containerName="registry-server" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.096247 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.099197 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-vs45x" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.104605 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz"] Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.164024 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c19dba91-7b40-4824-9377-1e55368da0c7-util\") pod \"bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz\" (UID: \"c19dba91-7b40-4824-9377-1e55368da0c7\") " pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.164092 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c19dba91-7b40-4824-9377-1e55368da0c7-bundle\") pod \"bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz\" (UID: \"c19dba91-7b40-4824-9377-1e55368da0c7\") " pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.164130 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsph5\" (UniqueName: \"kubernetes.io/projected/c19dba91-7b40-4824-9377-1e55368da0c7-kube-api-access-xsph5\") pod \"bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz\" (UID: \"c19dba91-7b40-4824-9377-1e55368da0c7\") " pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.265744 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c19dba91-7b40-4824-9377-1e55368da0c7-util\") pod \"bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz\" (UID: \"c19dba91-7b40-4824-9377-1e55368da0c7\") " pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.265871 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c19dba91-7b40-4824-9377-1e55368da0c7-bundle\") pod \"bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz\" (UID: \"c19dba91-7b40-4824-9377-1e55368da0c7\") " pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.265948 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsph5\" (UniqueName: \"kubernetes.io/projected/c19dba91-7b40-4824-9377-1e55368da0c7-kube-api-access-xsph5\") pod \"bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz\" (UID: \"c19dba91-7b40-4824-9377-1e55368da0c7\") " pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.267974 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c19dba91-7b40-4824-9377-1e55368da0c7-bundle\") pod \"bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz\" (UID: \"c19dba91-7b40-4824-9377-1e55368da0c7\") " pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.268049 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c19dba91-7b40-4824-9377-1e55368da0c7-util\") pod \"bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz\" (UID: \"c19dba91-7b40-4824-9377-1e55368da0c7\") " pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.290799 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsph5\" (UniqueName: \"kubernetes.io/projected/c19dba91-7b40-4824-9377-1e55368da0c7-kube-api-access-xsph5\") pod \"bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz\" (UID: \"c19dba91-7b40-4824-9377-1e55368da0c7\") " pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.421904 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.653116 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz"] Jan 26 09:19:21 crc kubenswrapper[4872]: W0126 09:19:21.660071 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc19dba91_7b40_4824_9377_1e55368da0c7.slice/crio-2b3d8497d08204f0c42429f8296c15011e101cb72edfe6ea42b77d16ce8fe2d4 WatchSource:0}: Error finding container 2b3d8497d08204f0c42429f8296c15011e101cb72edfe6ea42b77d16ce8fe2d4: Status 404 returned error can't find the container with id 2b3d8497d08204f0c42429f8296c15011e101cb72edfe6ea42b77d16ce8fe2d4 Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.849156 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" event={"ID":"c19dba91-7b40-4824-9377-1e55368da0c7","Type":"ContainerStarted","Data":"f25feef069781b7f46937230975eae6d0b409fbce6bea6935343aba276d6d390"} Jan 26 09:19:21 crc kubenswrapper[4872]: I0126 09:19:21.849222 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" event={"ID":"c19dba91-7b40-4824-9377-1e55368da0c7","Type":"ContainerStarted","Data":"2b3d8497d08204f0c42429f8296c15011e101cb72edfe6ea42b77d16ce8fe2d4"} Jan 26 09:19:22 crc kubenswrapper[4872]: I0126 09:19:22.862191 4872 generic.go:334] "Generic (PLEG): container finished" podID="c19dba91-7b40-4824-9377-1e55368da0c7" containerID="f25feef069781b7f46937230975eae6d0b409fbce6bea6935343aba276d6d390" exitCode=0 Jan 26 09:19:22 crc kubenswrapper[4872]: I0126 09:19:22.862258 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" event={"ID":"c19dba91-7b40-4824-9377-1e55368da0c7","Type":"ContainerDied","Data":"f25feef069781b7f46937230975eae6d0b409fbce6bea6935343aba276d6d390"} Jan 26 09:19:23 crc kubenswrapper[4872]: I0126 09:19:23.872926 4872 generic.go:334] "Generic (PLEG): container finished" podID="c19dba91-7b40-4824-9377-1e55368da0c7" containerID="195622fb74e3c6ee23e3bd1393981d421bdabb48987874f53f9d30762388cf66" exitCode=0 Jan 26 09:19:23 crc kubenswrapper[4872]: I0126 09:19:23.873011 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" event={"ID":"c19dba91-7b40-4824-9377-1e55368da0c7","Type":"ContainerDied","Data":"195622fb74e3c6ee23e3bd1393981d421bdabb48987874f53f9d30762388cf66"} Jan 26 09:19:24 crc kubenswrapper[4872]: I0126 09:19:24.885004 4872 generic.go:334] "Generic (PLEG): container finished" podID="c19dba91-7b40-4824-9377-1e55368da0c7" containerID="c224883e4e9bcc43eff522a8fc6d86849ee0ec0c4c58b011023c60defedc3073" exitCode=0 Jan 26 09:19:24 crc kubenswrapper[4872]: I0126 09:19:24.885150 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" event={"ID":"c19dba91-7b40-4824-9377-1e55368da0c7","Type":"ContainerDied","Data":"c224883e4e9bcc43eff522a8fc6d86849ee0ec0c4c58b011023c60defedc3073"} Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.200358 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.355057 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsph5\" (UniqueName: \"kubernetes.io/projected/c19dba91-7b40-4824-9377-1e55368da0c7-kube-api-access-xsph5\") pod \"c19dba91-7b40-4824-9377-1e55368da0c7\" (UID: \"c19dba91-7b40-4824-9377-1e55368da0c7\") " Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.355280 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c19dba91-7b40-4824-9377-1e55368da0c7-bundle\") pod \"c19dba91-7b40-4824-9377-1e55368da0c7\" (UID: \"c19dba91-7b40-4824-9377-1e55368da0c7\") " Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.355357 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c19dba91-7b40-4824-9377-1e55368da0c7-util\") pod \"c19dba91-7b40-4824-9377-1e55368da0c7\" (UID: \"c19dba91-7b40-4824-9377-1e55368da0c7\") " Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.356189 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c19dba91-7b40-4824-9377-1e55368da0c7-bundle" (OuterVolumeSpecName: "bundle") pod "c19dba91-7b40-4824-9377-1e55368da0c7" (UID: "c19dba91-7b40-4824-9377-1e55368da0c7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.362474 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c19dba91-7b40-4824-9377-1e55368da0c7-kube-api-access-xsph5" (OuterVolumeSpecName: "kube-api-access-xsph5") pod "c19dba91-7b40-4824-9377-1e55368da0c7" (UID: "c19dba91-7b40-4824-9377-1e55368da0c7"). InnerVolumeSpecName "kube-api-access-xsph5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.369883 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c19dba91-7b40-4824-9377-1e55368da0c7-util" (OuterVolumeSpecName: "util") pod "c19dba91-7b40-4824-9377-1e55368da0c7" (UID: "c19dba91-7b40-4824-9377-1e55368da0c7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.457268 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsph5\" (UniqueName: \"kubernetes.io/projected/c19dba91-7b40-4824-9377-1e55368da0c7-kube-api-access-xsph5\") on node \"crc\" DevicePath \"\"" Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.457320 4872 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c19dba91-7b40-4824-9377-1e55368da0c7-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.457340 4872 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c19dba91-7b40-4824-9377-1e55368da0c7-util\") on node \"crc\" DevicePath \"\"" Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.906872 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" event={"ID":"c19dba91-7b40-4824-9377-1e55368da0c7","Type":"ContainerDied","Data":"2b3d8497d08204f0c42429f8296c15011e101cb72edfe6ea42b77d16ce8fe2d4"} Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.907370 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b3d8497d08204f0c42429f8296c15011e101cb72edfe6ea42b77d16ce8fe2d4" Jan 26 09:19:26 crc kubenswrapper[4872]: I0126 09:19:26.906971 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz" Jan 26 09:19:33 crc kubenswrapper[4872]: I0126 09:19:33.473764 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28"] Jan 26 09:19:33 crc kubenswrapper[4872]: E0126 09:19:33.474873 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19dba91-7b40-4824-9377-1e55368da0c7" containerName="util" Jan 26 09:19:33 crc kubenswrapper[4872]: I0126 09:19:33.474889 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19dba91-7b40-4824-9377-1e55368da0c7" containerName="util" Jan 26 09:19:33 crc kubenswrapper[4872]: E0126 09:19:33.474927 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19dba91-7b40-4824-9377-1e55368da0c7" containerName="pull" Jan 26 09:19:33 crc kubenswrapper[4872]: I0126 09:19:33.474941 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19dba91-7b40-4824-9377-1e55368da0c7" containerName="pull" Jan 26 09:19:33 crc kubenswrapper[4872]: E0126 09:19:33.474953 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19dba91-7b40-4824-9377-1e55368da0c7" containerName="extract" Jan 26 09:19:33 crc kubenswrapper[4872]: I0126 09:19:33.474961 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19dba91-7b40-4824-9377-1e55368da0c7" containerName="extract" Jan 26 09:19:33 crc kubenswrapper[4872]: I0126 09:19:33.475080 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="c19dba91-7b40-4824-9377-1e55368da0c7" containerName="extract" Jan 26 09:19:33 crc kubenswrapper[4872]: I0126 09:19:33.475592 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28" Jan 26 09:19:33 crc kubenswrapper[4872]: I0126 09:19:33.478921 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-57qr9" Jan 26 09:19:33 crc kubenswrapper[4872]: I0126 09:19:33.505722 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28"] Jan 26 09:19:33 crc kubenswrapper[4872]: I0126 09:19:33.557209 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxfm7\" (UniqueName: \"kubernetes.io/projected/8582c505-0f00-4751-bae0-b9324821e2cb-kube-api-access-xxfm7\") pod \"openstack-operator-controller-init-8578c7d94c-gfq28\" (UID: \"8582c505-0f00-4751-bae0-b9324821e2cb\") " pod="openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28" Jan 26 09:19:33 crc kubenswrapper[4872]: I0126 09:19:33.658232 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxfm7\" (UniqueName: \"kubernetes.io/projected/8582c505-0f00-4751-bae0-b9324821e2cb-kube-api-access-xxfm7\") pod \"openstack-operator-controller-init-8578c7d94c-gfq28\" (UID: \"8582c505-0f00-4751-bae0-b9324821e2cb\") " pod="openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28" Jan 26 09:19:33 crc kubenswrapper[4872]: I0126 09:19:33.678495 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxfm7\" (UniqueName: \"kubernetes.io/projected/8582c505-0f00-4751-bae0-b9324821e2cb-kube-api-access-xxfm7\") pod \"openstack-operator-controller-init-8578c7d94c-gfq28\" (UID: \"8582c505-0f00-4751-bae0-b9324821e2cb\") " pod="openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28" Jan 26 09:19:33 crc kubenswrapper[4872]: I0126 09:19:33.792642 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28" Jan 26 09:19:34 crc kubenswrapper[4872]: I0126 09:19:34.054525 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28"] Jan 26 09:19:35 crc kubenswrapper[4872]: I0126 09:19:35.079363 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28" event={"ID":"8582c505-0f00-4751-bae0-b9324821e2cb","Type":"ContainerStarted","Data":"7809bf91d13fd6973bca1c82debd32ef8c23cb9425516a076756b20873839fb5"} Jan 26 09:19:40 crc kubenswrapper[4872]: I0126 09:19:40.125413 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28" event={"ID":"8582c505-0f00-4751-bae0-b9324821e2cb","Type":"ContainerStarted","Data":"1a934c1026d6bd2c861b2d86d6dc0e4edfb566dbff73633ee09ca3880b522a2d"} Jan 26 09:19:40 crc kubenswrapper[4872]: I0126 09:19:40.126503 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28" Jan 26 09:19:40 crc kubenswrapper[4872]: I0126 09:19:40.166739 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28" podStartSLOduration=1.828857891 podStartE2EDuration="7.166700555s" podCreationTimestamp="2026-01-26 09:19:33 +0000 UTC" firstStartedPulling="2026-01-26 09:19:34.069742445 +0000 UTC m=+707.378582236" lastFinishedPulling="2026-01-26 09:19:39.407585099 +0000 UTC m=+712.716424900" observedRunningTime="2026-01-26 09:19:40.161021556 +0000 UTC m=+713.469861437" watchObservedRunningTime="2026-01-26 09:19:40.166700555 +0000 UTC m=+713.475540386" Jan 26 09:19:40 crc kubenswrapper[4872]: I0126 09:19:40.228369 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:19:40 crc kubenswrapper[4872]: I0126 09:19:40.228606 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:19:43 crc kubenswrapper[4872]: I0126 09:19:43.003245 4872 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Jan 26 09:19:51 crc kubenswrapper[4872]: I0126 09:19:51.984020 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zt4dl"] Jan 26 09:19:51 crc kubenswrapper[4872]: I0126 09:19:51.986051 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:19:51 crc kubenswrapper[4872]: I0126 09:19:51.997550 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zt4dl"] Jan 26 09:19:52 crc kubenswrapper[4872]: I0126 09:19:52.123759 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/819be2d5-2d70-458e-8326-d7e942ce4e02-utilities\") pod \"community-operators-zt4dl\" (UID: \"819be2d5-2d70-458e-8326-d7e942ce4e02\") " pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:19:52 crc kubenswrapper[4872]: I0126 09:19:52.123872 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/819be2d5-2d70-458e-8326-d7e942ce4e02-catalog-content\") pod \"community-operators-zt4dl\" (UID: \"819be2d5-2d70-458e-8326-d7e942ce4e02\") " pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:19:52 crc kubenswrapper[4872]: I0126 09:19:52.123960 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mhzq\" (UniqueName: \"kubernetes.io/projected/819be2d5-2d70-458e-8326-d7e942ce4e02-kube-api-access-9mhzq\") pod \"community-operators-zt4dl\" (UID: \"819be2d5-2d70-458e-8326-d7e942ce4e02\") " pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:19:52 crc kubenswrapper[4872]: I0126 09:19:52.225257 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mhzq\" (UniqueName: \"kubernetes.io/projected/819be2d5-2d70-458e-8326-d7e942ce4e02-kube-api-access-9mhzq\") pod \"community-operators-zt4dl\" (UID: \"819be2d5-2d70-458e-8326-d7e942ce4e02\") " pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:19:52 crc kubenswrapper[4872]: I0126 09:19:52.225332 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/819be2d5-2d70-458e-8326-d7e942ce4e02-utilities\") pod \"community-operators-zt4dl\" (UID: \"819be2d5-2d70-458e-8326-d7e942ce4e02\") " pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:19:52 crc kubenswrapper[4872]: I0126 09:19:52.225374 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/819be2d5-2d70-458e-8326-d7e942ce4e02-catalog-content\") pod \"community-operators-zt4dl\" (UID: \"819be2d5-2d70-458e-8326-d7e942ce4e02\") " pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:19:52 crc kubenswrapper[4872]: I0126 09:19:52.226019 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/819be2d5-2d70-458e-8326-d7e942ce4e02-utilities\") pod \"community-operators-zt4dl\" (UID: \"819be2d5-2d70-458e-8326-d7e942ce4e02\") " pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:19:52 crc kubenswrapper[4872]: I0126 09:19:52.226140 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/819be2d5-2d70-458e-8326-d7e942ce4e02-catalog-content\") pod \"community-operators-zt4dl\" (UID: \"819be2d5-2d70-458e-8326-d7e942ce4e02\") " pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:19:52 crc kubenswrapper[4872]: I0126 09:19:52.261546 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mhzq\" (UniqueName: \"kubernetes.io/projected/819be2d5-2d70-458e-8326-d7e942ce4e02-kube-api-access-9mhzq\") pod \"community-operators-zt4dl\" (UID: \"819be2d5-2d70-458e-8326-d7e942ce4e02\") " pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:19:52 crc kubenswrapper[4872]: I0126 09:19:52.312785 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:19:52 crc kubenswrapper[4872]: I0126 09:19:52.614561 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zt4dl"] Jan 26 09:19:53 crc kubenswrapper[4872]: I0126 09:19:53.235908 4872 generic.go:334] "Generic (PLEG): container finished" podID="819be2d5-2d70-458e-8326-d7e942ce4e02" containerID="ae3d4ff416ab2c5ec602f51cd18a88aeed7036737cc36ba301033e7ffef626bc" exitCode=0 Jan 26 09:19:53 crc kubenswrapper[4872]: I0126 09:19:53.236018 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zt4dl" event={"ID":"819be2d5-2d70-458e-8326-d7e942ce4e02","Type":"ContainerDied","Data":"ae3d4ff416ab2c5ec602f51cd18a88aeed7036737cc36ba301033e7ffef626bc"} Jan 26 09:19:53 crc kubenswrapper[4872]: I0126 09:19:53.236320 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zt4dl" event={"ID":"819be2d5-2d70-458e-8326-d7e942ce4e02","Type":"ContainerStarted","Data":"e97f313323ebc0b3e3e287fbf06a9bfc6dc0bbc24e0f2cee7dc84e82ffb061b0"} Jan 26 09:19:53 crc kubenswrapper[4872]: I0126 09:19:53.797626 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-8578c7d94c-gfq28" Jan 26 09:19:54 crc kubenswrapper[4872]: I0126 09:19:54.243812 4872 generic.go:334] "Generic (PLEG): container finished" podID="819be2d5-2d70-458e-8326-d7e942ce4e02" containerID="aade5bd155afc9e34ac37e0cbacd80ca9abddcca71a775fa34adac74d45224dc" exitCode=0 Jan 26 09:19:54 crc kubenswrapper[4872]: I0126 09:19:54.244280 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zt4dl" event={"ID":"819be2d5-2d70-458e-8326-d7e942ce4e02","Type":"ContainerDied","Data":"aade5bd155afc9e34ac37e0cbacd80ca9abddcca71a775fa34adac74d45224dc"} Jan 26 09:19:55 crc kubenswrapper[4872]: I0126 09:19:55.255597 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zt4dl" event={"ID":"819be2d5-2d70-458e-8326-d7e942ce4e02","Type":"ContainerStarted","Data":"cfc1f8188732b4250c3d1c78b1b3bf18eab3fcb41f5dcb0937c216e78d41f78f"} Jan 26 09:19:55 crc kubenswrapper[4872]: I0126 09:19:55.275538 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zt4dl" podStartSLOduration=2.833114034 podStartE2EDuration="4.275515834s" podCreationTimestamp="2026-01-26 09:19:51 +0000 UTC" firstStartedPulling="2026-01-26 09:19:53.239305397 +0000 UTC m=+726.548145198" lastFinishedPulling="2026-01-26 09:19:54.681707197 +0000 UTC m=+727.990546998" observedRunningTime="2026-01-26 09:19:55.273561574 +0000 UTC m=+728.582401375" watchObservedRunningTime="2026-01-26 09:19:55.275515834 +0000 UTC m=+728.584355635" Jan 26 09:20:02 crc kubenswrapper[4872]: I0126 09:20:02.313373 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:20:02 crc kubenswrapper[4872]: I0126 09:20:02.314265 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:20:02 crc kubenswrapper[4872]: I0126 09:20:02.373378 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:20:03 crc kubenswrapper[4872]: I0126 09:20:03.371959 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:20:03 crc kubenswrapper[4872]: I0126 09:20:03.432527 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zt4dl"] Jan 26 09:20:05 crc kubenswrapper[4872]: I0126 09:20:05.331518 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zt4dl" podUID="819be2d5-2d70-458e-8326-d7e942ce4e02" containerName="registry-server" containerID="cri-o://cfc1f8188732b4250c3d1c78b1b3bf18eab3fcb41f5dcb0937c216e78d41f78f" gracePeriod=2 Jan 26 09:20:06 crc kubenswrapper[4872]: I0126 09:20:06.342292 4872 generic.go:334] "Generic (PLEG): container finished" podID="819be2d5-2d70-458e-8326-d7e942ce4e02" containerID="cfc1f8188732b4250c3d1c78b1b3bf18eab3fcb41f5dcb0937c216e78d41f78f" exitCode=0 Jan 26 09:20:06 crc kubenswrapper[4872]: I0126 09:20:06.342379 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zt4dl" event={"ID":"819be2d5-2d70-458e-8326-d7e942ce4e02","Type":"ContainerDied","Data":"cfc1f8188732b4250c3d1c78b1b3bf18eab3fcb41f5dcb0937c216e78d41f78f"} Jan 26 09:20:06 crc kubenswrapper[4872]: I0126 09:20:06.791702 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:20:06 crc kubenswrapper[4872]: I0126 09:20:06.977481 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/819be2d5-2d70-458e-8326-d7e942ce4e02-utilities\") pod \"819be2d5-2d70-458e-8326-d7e942ce4e02\" (UID: \"819be2d5-2d70-458e-8326-d7e942ce4e02\") " Jan 26 09:20:06 crc kubenswrapper[4872]: I0126 09:20:06.977650 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mhzq\" (UniqueName: \"kubernetes.io/projected/819be2d5-2d70-458e-8326-d7e942ce4e02-kube-api-access-9mhzq\") pod \"819be2d5-2d70-458e-8326-d7e942ce4e02\" (UID: \"819be2d5-2d70-458e-8326-d7e942ce4e02\") " Jan 26 09:20:06 crc kubenswrapper[4872]: I0126 09:20:06.977713 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/819be2d5-2d70-458e-8326-d7e942ce4e02-catalog-content\") pod \"819be2d5-2d70-458e-8326-d7e942ce4e02\" (UID: \"819be2d5-2d70-458e-8326-d7e942ce4e02\") " Jan 26 09:20:06 crc kubenswrapper[4872]: I0126 09:20:06.979152 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/819be2d5-2d70-458e-8326-d7e942ce4e02-utilities" (OuterVolumeSpecName: "utilities") pod "819be2d5-2d70-458e-8326-d7e942ce4e02" (UID: "819be2d5-2d70-458e-8326-d7e942ce4e02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:20:06 crc kubenswrapper[4872]: I0126 09:20:06.991187 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/819be2d5-2d70-458e-8326-d7e942ce4e02-kube-api-access-9mhzq" (OuterVolumeSpecName: "kube-api-access-9mhzq") pod "819be2d5-2d70-458e-8326-d7e942ce4e02" (UID: "819be2d5-2d70-458e-8326-d7e942ce4e02"). InnerVolumeSpecName "kube-api-access-9mhzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:20:07 crc kubenswrapper[4872]: I0126 09:20:07.060576 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/819be2d5-2d70-458e-8326-d7e942ce4e02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "819be2d5-2d70-458e-8326-d7e942ce4e02" (UID: "819be2d5-2d70-458e-8326-d7e942ce4e02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:20:07 crc kubenswrapper[4872]: I0126 09:20:07.080952 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/819be2d5-2d70-458e-8326-d7e942ce4e02-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:20:07 crc kubenswrapper[4872]: I0126 09:20:07.080993 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mhzq\" (UniqueName: \"kubernetes.io/projected/819be2d5-2d70-458e-8326-d7e942ce4e02-kube-api-access-9mhzq\") on node \"crc\" DevicePath \"\"" Jan 26 09:20:07 crc kubenswrapper[4872]: I0126 09:20:07.081006 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/819be2d5-2d70-458e-8326-d7e942ce4e02-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:20:07 crc kubenswrapper[4872]: I0126 09:20:07.370176 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zt4dl" event={"ID":"819be2d5-2d70-458e-8326-d7e942ce4e02","Type":"ContainerDied","Data":"e97f313323ebc0b3e3e287fbf06a9bfc6dc0bbc24e0f2cee7dc84e82ffb061b0"} Jan 26 09:20:07 crc kubenswrapper[4872]: I0126 09:20:07.370244 4872 scope.go:117] "RemoveContainer" containerID="cfc1f8188732b4250c3d1c78b1b3bf18eab3fcb41f5dcb0937c216e78d41f78f" Jan 26 09:20:07 crc kubenswrapper[4872]: I0126 09:20:07.370275 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zt4dl" Jan 26 09:20:07 crc kubenswrapper[4872]: I0126 09:20:07.390746 4872 scope.go:117] "RemoveContainer" containerID="aade5bd155afc9e34ac37e0cbacd80ca9abddcca71a775fa34adac74d45224dc" Jan 26 09:20:07 crc kubenswrapper[4872]: I0126 09:20:07.434092 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zt4dl"] Jan 26 09:20:07 crc kubenswrapper[4872]: I0126 09:20:07.446468 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zt4dl"] Jan 26 09:20:07 crc kubenswrapper[4872]: I0126 09:20:07.448998 4872 scope.go:117] "RemoveContainer" containerID="ae3d4ff416ab2c5ec602f51cd18a88aeed7036737cc36ba301033e7ffef626bc" Jan 26 09:20:09 crc kubenswrapper[4872]: I0126 09:20:09.220224 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="819be2d5-2d70-458e-8326-d7e942ce4e02" path="/var/lib/kubelet/pods/819be2d5-2d70-458e-8326-d7e942ce4e02/volumes" Jan 26 09:20:10 crc kubenswrapper[4872]: I0126 09:20:10.227522 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:20:10 crc kubenswrapper[4872]: I0126 09:20:10.227610 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.194046 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757"] Jan 26 09:20:21 crc kubenswrapper[4872]: E0126 09:20:21.195155 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="819be2d5-2d70-458e-8326-d7e942ce4e02" containerName="registry-server" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.195172 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="819be2d5-2d70-458e-8326-d7e942ce4e02" containerName="registry-server" Jan 26 09:20:21 crc kubenswrapper[4872]: E0126 09:20:21.195200 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="819be2d5-2d70-458e-8326-d7e942ce4e02" containerName="extract-content" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.195207 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="819be2d5-2d70-458e-8326-d7e942ce4e02" containerName="extract-content" Jan 26 09:20:21 crc kubenswrapper[4872]: E0126 09:20:21.195217 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="819be2d5-2d70-458e-8326-d7e942ce4e02" containerName="extract-utilities" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.195226 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="819be2d5-2d70-458e-8326-d7e942ce4e02" containerName="extract-utilities" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.195348 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="819be2d5-2d70-458e-8326-d7e942ce4e02" containerName="registry-server" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.195896 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.198771 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-ntrkf" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.202369 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.210486 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.214568 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.216463 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-nmh2d" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.218231 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.219159 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.222211 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-8bbld" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.239147 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqhbj\" (UniqueName: \"kubernetes.io/projected/866888fd-6bb1-4184-9e17-352cb8d3727f-kube-api-access-pqhbj\") pod \"barbican-operator-controller-manager-7f86f8796f-96757\" (UID: \"866888fd-6bb1-4184-9e17-352cb8d3727f\") " pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.253273 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.254369 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.256981 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-5glfq" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.311968 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.317037 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.318400 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.323084 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-z4stm" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.325171 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.326088 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.328323 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-zc9nf" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.341319 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-758868c854-rss8g"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.342472 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxd8z\" (UniqueName: \"kubernetes.io/projected/a313f841-7d48-4621-a385-b001d8cfa11b-kube-api-access-cxd8z\") pod \"designate-operator-controller-manager-b45d7bf98-ldqqh\" (UID: \"a313f841-7d48-4621-a385-b001d8cfa11b\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.342552 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqhbj\" (UniqueName: \"kubernetes.io/projected/866888fd-6bb1-4184-9e17-352cb8d3727f-kube-api-access-pqhbj\") pod \"barbican-operator-controller-manager-7f86f8796f-96757\" (UID: \"866888fd-6bb1-4184-9e17-352cb8d3727f\") " pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.342582 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lgmh\" (UniqueName: \"kubernetes.io/projected/b3751f8d-5921-4b5f-b740-1ba5fc532f7b-kube-api-access-9lgmh\") pod \"cinder-operator-controller-manager-7478f7dbf9-cvcnq\" (UID: \"b3751f8d-5921-4b5f-b740-1ba5fc532f7b\") " pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.342611 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv9nh\" (UniqueName: \"kubernetes.io/projected/7f35eab0-a711-4b5d-8b5a-05c72414c30c-kube-api-access-lv9nh\") pod \"glance-operator-controller-manager-78fdd796fd-cwkcb\" (UID: \"7f35eab0-a711-4b5d-8b5a-05c72414c30c\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.342917 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.346122 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-ts4ww" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.346148 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.355451 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.356829 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.361219 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-922hp" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.375301 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.376284 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.379770 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqhbj\" (UniqueName: \"kubernetes.io/projected/866888fd-6bb1-4184-9e17-352cb8d3727f-kube-api-access-pqhbj\") pod \"barbican-operator-controller-manager-7f86f8796f-96757\" (UID: \"866888fd-6bb1-4184-9e17-352cb8d3727f\") " pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.387737 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-dq94z" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.391052 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.445103 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert\") pod \"infra-operator-controller-manager-758868c854-rss8g\" (UID: \"5619ad3b-5066-4163-9cff-7654030fc5b1\") " pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.445176 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pln5b\" (UniqueName: \"kubernetes.io/projected/c5bfadfb-1ac5-475e-8b5f-8096dd00d850-kube-api-access-pln5b\") pod \"heat-operator-controller-manager-594c8c9d5d-g5txh\" (UID: \"c5bfadfb-1ac5-475e-8b5f-8096dd00d850\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.445221 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbglf\" (UniqueName: \"kubernetes.io/projected/87e00bbc-66f4-4921-94ce-24d7b658341c-kube-api-access-bbglf\") pod \"keystone-operator-controller-manager-b8b6d4659-q4kjt\" (UID: \"87e00bbc-66f4-4921-94ce-24d7b658341c\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.445274 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxd8z\" (UniqueName: \"kubernetes.io/projected/a313f841-7d48-4621-a385-b001d8cfa11b-kube-api-access-cxd8z\") pod \"designate-operator-controller-manager-b45d7bf98-ldqqh\" (UID: \"a313f841-7d48-4621-a385-b001d8cfa11b\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.445324 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km4cd\" (UniqueName: \"kubernetes.io/projected/9cdfe7c6-5b86-4204-b9bf-3597361af9fe-kube-api-access-km4cd\") pod \"ironic-operator-controller-manager-6b46765b57-crcq2\" (UID: \"9cdfe7c6-5b86-4204-b9bf-3597361af9fe\") " pod="openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.445350 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lgmh\" (UniqueName: \"kubernetes.io/projected/b3751f8d-5921-4b5f-b740-1ba5fc532f7b-kube-api-access-9lgmh\") pod \"cinder-operator-controller-manager-7478f7dbf9-cvcnq\" (UID: \"b3751f8d-5921-4b5f-b740-1ba5fc532f7b\") " pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.445407 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv9nh\" (UniqueName: \"kubernetes.io/projected/7f35eab0-a711-4b5d-8b5a-05c72414c30c-kube-api-access-lv9nh\") pod \"glance-operator-controller-manager-78fdd796fd-cwkcb\" (UID: \"7f35eab0-a711-4b5d-8b5a-05c72414c30c\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.446023 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s7xd\" (UniqueName: \"kubernetes.io/projected/2bda61c9-b3dd-476a-86fb-882a5f172ba9-kube-api-access-9s7xd\") pod \"horizon-operator-controller-manager-77d5c5b54f-s4b5m\" (UID: \"2bda61c9-b3dd-476a-86fb-882a5f172ba9\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.446155 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2mq5\" (UniqueName: \"kubernetes.io/projected/5619ad3b-5066-4163-9cff-7654030fc5b1-kube-api-access-z2mq5\") pod \"infra-operator-controller-manager-758868c854-rss8g\" (UID: \"5619ad3b-5066-4163-9cff-7654030fc5b1\") " pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.446870 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.487191 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv9nh\" (UniqueName: \"kubernetes.io/projected/7f35eab0-a711-4b5d-8b5a-05c72414c30c-kube-api-access-lv9nh\") pod \"glance-operator-controller-manager-78fdd796fd-cwkcb\" (UID: \"7f35eab0-a711-4b5d-8b5a-05c72414c30c\") " pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.489120 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxd8z\" (UniqueName: \"kubernetes.io/projected/a313f841-7d48-4621-a385-b001d8cfa11b-kube-api-access-cxd8z\") pod \"designate-operator-controller-manager-b45d7bf98-ldqqh\" (UID: \"a313f841-7d48-4621-a385-b001d8cfa11b\") " pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.511642 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.513918 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lgmh\" (UniqueName: \"kubernetes.io/projected/b3751f8d-5921-4b5f-b740-1ba5fc532f7b-kube-api-access-9lgmh\") pod \"cinder-operator-controller-manager-7478f7dbf9-cvcnq\" (UID: \"b3751f8d-5921-4b5f-b740-1ba5fc532f7b\") " pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.517391 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.521989 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-758868c854-rss8g"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.527021 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.536425 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.547939 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s7xd\" (UniqueName: \"kubernetes.io/projected/2bda61c9-b3dd-476a-86fb-882a5f172ba9-kube-api-access-9s7xd\") pod \"horizon-operator-controller-manager-77d5c5b54f-s4b5m\" (UID: \"2bda61c9-b3dd-476a-86fb-882a5f172ba9\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.547997 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2mq5\" (UniqueName: \"kubernetes.io/projected/5619ad3b-5066-4163-9cff-7654030fc5b1-kube-api-access-z2mq5\") pod \"infra-operator-controller-manager-758868c854-rss8g\" (UID: \"5619ad3b-5066-4163-9cff-7654030fc5b1\") " pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.548033 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert\") pod \"infra-operator-controller-manager-758868c854-rss8g\" (UID: \"5619ad3b-5066-4163-9cff-7654030fc5b1\") " pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.548063 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pln5b\" (UniqueName: \"kubernetes.io/projected/c5bfadfb-1ac5-475e-8b5f-8096dd00d850-kube-api-access-pln5b\") pod \"heat-operator-controller-manager-594c8c9d5d-g5txh\" (UID: \"c5bfadfb-1ac5-475e-8b5f-8096dd00d850\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.548083 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbglf\" (UniqueName: \"kubernetes.io/projected/87e00bbc-66f4-4921-94ce-24d7b658341c-kube-api-access-bbglf\") pod \"keystone-operator-controller-manager-b8b6d4659-q4kjt\" (UID: \"87e00bbc-66f4-4921-94ce-24d7b658341c\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.548124 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km4cd\" (UniqueName: \"kubernetes.io/projected/9cdfe7c6-5b86-4204-b9bf-3597361af9fe-kube-api-access-km4cd\") pod \"ironic-operator-controller-manager-6b46765b57-crcq2\" (UID: \"9cdfe7c6-5b86-4204-b9bf-3597361af9fe\") " pod="openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2" Jan 26 09:20:21 crc kubenswrapper[4872]: E0126 09:20:21.548772 4872 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 26 09:20:21 crc kubenswrapper[4872]: E0126 09:20:21.548849 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert podName:5619ad3b-5066-4163-9cff-7654030fc5b1 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:22.04883254 +0000 UTC m=+755.357672341 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert") pod "infra-operator-controller-manager-758868c854-rss8g" (UID: "5619ad3b-5066-4163-9cff-7654030fc5b1") : secret "infra-operator-webhook-server-cert" not found Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.549238 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.575928 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.576016 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.577643 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2mq5\" (UniqueName: \"kubernetes.io/projected/5619ad3b-5066-4163-9cff-7654030fc5b1-kube-api-access-z2mq5\") pod \"infra-operator-controller-manager-758868c854-rss8g\" (UID: \"5619ad3b-5066-4163-9cff-7654030fc5b1\") " pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.578261 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.581707 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-7p9dn" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.581980 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.583241 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.589529 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pln5b\" (UniqueName: \"kubernetes.io/projected/c5bfadfb-1ac5-475e-8b5f-8096dd00d850-kube-api-access-pln5b\") pod \"heat-operator-controller-manager-594c8c9d5d-g5txh\" (UID: \"c5bfadfb-1ac5-475e-8b5f-8096dd00d850\") " pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.589528 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.591618 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.596027 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-qgdhr" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.596954 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbglf\" (UniqueName: \"kubernetes.io/projected/87e00bbc-66f4-4921-94ce-24d7b658341c-kube-api-access-bbglf\") pod \"keystone-operator-controller-manager-b8b6d4659-q4kjt\" (UID: \"87e00bbc-66f4-4921-94ce-24d7b658341c\") " pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.599547 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km4cd\" (UniqueName: \"kubernetes.io/projected/9cdfe7c6-5b86-4204-b9bf-3597361af9fe-kube-api-access-km4cd\") pod \"ironic-operator-controller-manager-6b46765b57-crcq2\" (UID: \"9cdfe7c6-5b86-4204-b9bf-3597361af9fe\") " pod="openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.615875 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.617001 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.620070 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-7m5ht" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.620919 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s7xd\" (UniqueName: \"kubernetes.io/projected/2bda61c9-b3dd-476a-86fb-882a5f172ba9-kube-api-access-9s7xd\") pod \"horizon-operator-controller-manager-77d5c5b54f-s4b5m\" (UID: \"2bda61c9-b3dd-476a-86fb-882a5f172ba9\") " pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.635153 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.640790 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.645403 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.649283 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcwt2\" (UniqueName: \"kubernetes.io/projected/c05670e5-8312-46ce-9fe7-da7ecc691465-kube-api-access-mcwt2\") pod \"neutron-operator-controller-manager-78d58447c5-qwp9z\" (UID: \"c05670e5-8312-46ce-9fe7-da7ecc691465\") " pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.649547 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n44p8\" (UniqueName: \"kubernetes.io/projected/40ff0d90-ae83-47b6-b6b0-8969b47bcf92-kube-api-access-n44p8\") pod \"mariadb-operator-controller-manager-6b9fb5fdcb-tcc58\" (UID: \"40ff0d90-ae83-47b6-b6b0-8969b47bcf92\") " pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.649703 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvb78\" (UniqueName: \"kubernetes.io/projected/4d3db7ee-d472-427c-96a2-1c0f67a3b507-kube-api-access-gvb78\") pod \"manila-operator-controller-manager-78c6999f6f-wct5s\" (UID: \"4d3db7ee-d472-427c-96a2-1c0f67a3b507\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.651209 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.659946 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.665401 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.667044 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.667833 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.668603 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.671390 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-t9xd6" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.671931 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-r78pj" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.678633 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.693812 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.712598 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.715475 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.719150 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.720012 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.722541 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.723106 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.723154 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-fdhv2" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.724291 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-ppq9v" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.729706 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.731478 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.733940 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-7xpkr" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.740296 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.748166 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.751185 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcwt2\" (UniqueName: \"kubernetes.io/projected/c05670e5-8312-46ce-9fe7-da7ecc691465-kube-api-access-mcwt2\") pod \"neutron-operator-controller-manager-78d58447c5-qwp9z\" (UID: \"c05670e5-8312-46ce-9fe7-da7ecc691465\") " pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.751521 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqj9c\" (UniqueName: \"kubernetes.io/projected/acf08f03-b388-4e2a-99a2-2d492e490e67-kube-api-access-kqj9c\") pod \"nova-operator-controller-manager-7bdb645866-vz57m\" (UID: \"acf08f03-b388-4e2a-99a2-2d492e490e67\") " pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.751651 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv\" (UID: \"a0993a22-0883-4db0-9c07-c30657d9fb37\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.751760 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5qsw\" (UniqueName: \"kubernetes.io/projected/60fd216c-7ecc-477f-ac29-b7334ea5e450-kube-api-access-s5qsw\") pod \"placement-operator-controller-manager-79d5ccc684-7fm49\" (UID: \"60fd216c-7ecc-477f-ac29-b7334ea5e450\") " pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.751870 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n44p8\" (UniqueName: \"kubernetes.io/projected/40ff0d90-ae83-47b6-b6b0-8969b47bcf92-kube-api-access-n44p8\") pod \"mariadb-operator-controller-manager-6b9fb5fdcb-tcc58\" (UID: \"40ff0d90-ae83-47b6-b6b0-8969b47bcf92\") " pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.751955 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvb78\" (UniqueName: \"kubernetes.io/projected/4d3db7ee-d472-427c-96a2-1c0f67a3b507-kube-api-access-gvb78\") pod \"manila-operator-controller-manager-78c6999f6f-wct5s\" (UID: \"4d3db7ee-d472-427c-96a2-1c0f67a3b507\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.752048 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dbrq\" (UniqueName: \"kubernetes.io/projected/a0993a22-0883-4db0-9c07-c30657d9fb37-kube-api-access-5dbrq\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv\" (UID: \"a0993a22-0883-4db0-9c07-c30657d9fb37\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.752142 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vf26\" (UniqueName: \"kubernetes.io/projected/148da344-f5f5-43ab-ba3c-342f6cc10b47-kube-api-access-2vf26\") pod \"octavia-operator-controller-manager-5f4cd88d46-77t7p\" (UID: \"148da344-f5f5-43ab-ba3c-342f6cc10b47\") " pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.752223 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvbrp\" (UniqueName: \"kubernetes.io/projected/37007861-ed2f-4453-8a25-aa39f6b0f347-kube-api-access-bvbrp\") pod \"ovn-operator-controller-manager-6f75f45d54-r7rv2\" (UID: \"37007861-ed2f-4453-8a25-aa39f6b0f347\") " pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.753192 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.758999 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.764027 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.779278 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.779515 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.779825 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.780057 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n44p8\" (UniqueName: \"kubernetes.io/projected/40ff0d90-ae83-47b6-b6b0-8969b47bcf92-kube-api-access-n44p8\") pod \"mariadb-operator-controller-manager-6b9fb5fdcb-tcc58\" (UID: \"40ff0d90-ae83-47b6-b6b0-8969b47bcf92\") " pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.780165 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.782310 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-9k9kf" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.783378 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-vkm4z" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.788384 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.790215 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvb78\" (UniqueName: \"kubernetes.io/projected/4d3db7ee-d472-427c-96a2-1c0f67a3b507-kube-api-access-gvb78\") pod \"manila-operator-controller-manager-78c6999f6f-wct5s\" (UID: \"4d3db7ee-d472-427c-96a2-1c0f67a3b507\") " pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.794864 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcwt2\" (UniqueName: \"kubernetes.io/projected/c05670e5-8312-46ce-9fe7-da7ecc691465-kube-api-access-mcwt2\") pod \"neutron-operator-controller-manager-78d58447c5-qwp9z\" (UID: \"c05670e5-8312-46ce-9fe7-da7ecc691465\") " pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.828317 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.829949 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.832549 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-6ckx4" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.867482 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns"] Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.875242 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccbpt\" (UniqueName: \"kubernetes.io/projected/12d27df7-ee8b-4e48-b0fe-0b1f018e71f8-kube-api-access-ccbpt\") pod \"test-operator-controller-manager-69797bbcbd-tdnns\" (UID: \"12d27df7-ee8b-4e48-b0fe-0b1f018e71f8\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.875365 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vf26\" (UniqueName: \"kubernetes.io/projected/148da344-f5f5-43ab-ba3c-342f6cc10b47-kube-api-access-2vf26\") pod \"octavia-operator-controller-manager-5f4cd88d46-77t7p\" (UID: \"148da344-f5f5-43ab-ba3c-342f6cc10b47\") " pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.875444 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvbrp\" (UniqueName: \"kubernetes.io/projected/37007861-ed2f-4453-8a25-aa39f6b0f347-kube-api-access-bvbrp\") pod \"ovn-operator-controller-manager-6f75f45d54-r7rv2\" (UID: \"37007861-ed2f-4453-8a25-aa39f6b0f347\") " pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.875495 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws687\" (UniqueName: \"kubernetes.io/projected/d92c1258-d114-4696-a481-f9fff7faa046-kube-api-access-ws687\") pod \"swift-operator-controller-manager-547cbdb99f-cqr4g\" (UID: \"d92c1258-d114-4696-a481-f9fff7faa046\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.875575 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqj9c\" (UniqueName: \"kubernetes.io/projected/acf08f03-b388-4e2a-99a2-2d492e490e67-kube-api-access-kqj9c\") pod \"nova-operator-controller-manager-7bdb645866-vz57m\" (UID: \"acf08f03-b388-4e2a-99a2-2d492e490e67\") " pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.875621 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7pl7\" (UniqueName: \"kubernetes.io/projected/4b97d503-908a-48b0-bb0c-5e2cdcf251af-kube-api-access-x7pl7\") pod \"telemetry-operator-controller-manager-85cd9769bb-zfvrw\" (UID: \"4b97d503-908a-48b0-bb0c-5e2cdcf251af\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.875670 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv\" (UID: \"a0993a22-0883-4db0-9c07-c30657d9fb37\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.875738 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5qsw\" (UniqueName: \"kubernetes.io/projected/60fd216c-7ecc-477f-ac29-b7334ea5e450-kube-api-access-s5qsw\") pod \"placement-operator-controller-manager-79d5ccc684-7fm49\" (UID: \"60fd216c-7ecc-477f-ac29-b7334ea5e450\") " pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.875843 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dbrq\" (UniqueName: \"kubernetes.io/projected/a0993a22-0883-4db0-9c07-c30657d9fb37-kube-api-access-5dbrq\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv\" (UID: \"a0993a22-0883-4db0-9c07-c30657d9fb37\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:21 crc kubenswrapper[4872]: E0126 09:20:21.877705 4872 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 09:20:21 crc kubenswrapper[4872]: E0126 09:20:21.877954 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert podName:a0993a22-0883-4db0-9c07-c30657d9fb37 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:22.377745947 +0000 UTC m=+755.686585758 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" (UID: "a0993a22-0883-4db0-9c07-c30657d9fb37") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.938111 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5qsw\" (UniqueName: \"kubernetes.io/projected/60fd216c-7ecc-477f-ac29-b7334ea5e450-kube-api-access-s5qsw\") pod \"placement-operator-controller-manager-79d5ccc684-7fm49\" (UID: \"60fd216c-7ecc-477f-ac29-b7334ea5e450\") " pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.939192 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvbrp\" (UniqueName: \"kubernetes.io/projected/37007861-ed2f-4453-8a25-aa39f6b0f347-kube-api-access-bvbrp\") pod \"ovn-operator-controller-manager-6f75f45d54-r7rv2\" (UID: \"37007861-ed2f-4453-8a25-aa39f6b0f347\") " pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2" Jan 26 09:20:21 crc kubenswrapper[4872]: I0126 09:20:21.941321 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.011370 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.013445 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vf26\" (UniqueName: \"kubernetes.io/projected/148da344-f5f5-43ab-ba3c-342f6cc10b47-kube-api-access-2vf26\") pod \"octavia-operator-controller-manager-5f4cd88d46-77t7p\" (UID: \"148da344-f5f5-43ab-ba3c-342f6cc10b47\") " pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:21.967506 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqj9c\" (UniqueName: \"kubernetes.io/projected/acf08f03-b388-4e2a-99a2-2d492e490e67-kube-api-access-kqj9c\") pod \"nova-operator-controller-manager-7bdb645866-vz57m\" (UID: \"acf08f03-b388-4e2a-99a2-2d492e490e67\") " pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.015222 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dbrq\" (UniqueName: \"kubernetes.io/projected/a0993a22-0883-4db0-9c07-c30657d9fb37-kube-api-access-5dbrq\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv\" (UID: \"a0993a22-0883-4db0-9c07-c30657d9fb37\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.017522 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.020755 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.035655 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.091882 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-hfj2p"] Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.099163 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.099190 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7pl7\" (UniqueName: \"kubernetes.io/projected/4b97d503-908a-48b0-bb0c-5e2cdcf251af-kube-api-access-x7pl7\") pod \"telemetry-operator-controller-manager-85cd9769bb-zfvrw\" (UID: \"4b97d503-908a-48b0-bb0c-5e2cdcf251af\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.099427 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert\") pod \"infra-operator-controller-manager-758868c854-rss8g\" (UID: \"5619ad3b-5066-4163-9cff-7654030fc5b1\") " pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.099479 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccbpt\" (UniqueName: \"kubernetes.io/projected/12d27df7-ee8b-4e48-b0fe-0b1f018e71f8-kube-api-access-ccbpt\") pod \"test-operator-controller-manager-69797bbcbd-tdnns\" (UID: \"12d27df7-ee8b-4e48-b0fe-0b1f018e71f8\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.099655 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws687\" (UniqueName: \"kubernetes.io/projected/d92c1258-d114-4696-a481-f9fff7faa046-kube-api-access-ws687\") pod \"swift-operator-controller-manager-547cbdb99f-cqr4g\" (UID: \"d92c1258-d114-4696-a481-f9fff7faa046\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.099968 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.107032 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.108042 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-hfj2p"] Jan 26 09:20:22 crc kubenswrapper[4872]: E0126 09:20:22.109858 4872 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 26 09:20:22 crc kubenswrapper[4872]: E0126 09:20:22.109932 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert podName:5619ad3b-5066-4163-9cff-7654030fc5b1 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:23.109909151 +0000 UTC m=+756.418748952 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert") pod "infra-operator-controller-manager-758868c854-rss8g" (UID: "5619ad3b-5066-4163-9cff-7654030fc5b1") : secret "infra-operator-webhook-server-cert" not found Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.111829 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-6tgtj" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.153919 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws687\" (UniqueName: \"kubernetes.io/projected/d92c1258-d114-4696-a481-f9fff7faa046-kube-api-access-ws687\") pod \"swift-operator-controller-manager-547cbdb99f-cqr4g\" (UID: \"d92c1258-d114-4696-a481-f9fff7faa046\") " pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.155184 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccbpt\" (UniqueName: \"kubernetes.io/projected/12d27df7-ee8b-4e48-b0fe-0b1f018e71f8-kube-api-access-ccbpt\") pod \"test-operator-controller-manager-69797bbcbd-tdnns\" (UID: \"12d27df7-ee8b-4e48-b0fe-0b1f018e71f8\") " pod="openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.165948 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj"] Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.167556 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.173139 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7pl7\" (UniqueName: \"kubernetes.io/projected/4b97d503-908a-48b0-bb0c-5e2cdcf251af-kube-api-access-x7pl7\") pod \"telemetry-operator-controller-manager-85cd9769bb-zfvrw\" (UID: \"4b97d503-908a-48b0-bb0c-5e2cdcf251af\") " pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.174811 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.175052 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-l2482" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.175231 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.177310 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj"] Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.203876 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.204001 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.204069 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmb45\" (UniqueName: \"kubernetes.io/projected/6e3c0ac4-80fa-4d8d-864d-38056dc22402-kube-api-access-lmb45\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.204092 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vgsqx\" (UniqueName: \"kubernetes.io/projected/ced935ab-f3e5-470d-aa18-e15bb944fb86-kube-api-access-vgsqx\") pod \"watcher-operator-controller-manager-564965969-hfj2p\" (UID: \"ced935ab-f3e5-470d-aa18-e15bb944fb86\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.239334 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.257014 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.279988 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd"] Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.302421 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.305406 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmb45\" (UniqueName: \"kubernetes.io/projected/6e3c0ac4-80fa-4d8d-864d-38056dc22402-kube-api-access-lmb45\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.305471 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zxs6\" (UniqueName: \"kubernetes.io/projected/1fa3edb5-2354-4c32-ac85-d6be3a7dc16c-kube-api-access-5zxs6\") pod \"rabbitmq-cluster-operator-manager-668c99d594-hmfxd\" (UID: \"1fa3edb5-2354-4c32-ac85-d6be3a7dc16c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.305504 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vgsqx\" (UniqueName: \"kubernetes.io/projected/ced935ab-f3e5-470d-aa18-e15bb944fb86-kube-api-access-vgsqx\") pod \"watcher-operator-controller-manager-564965969-hfj2p\" (UID: \"ced935ab-f3e5-470d-aa18-e15bb944fb86\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.305554 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:22 crc kubenswrapper[4872]: E0126 09:20:22.305754 4872 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 26 09:20:22 crc kubenswrapper[4872]: E0126 09:20:22.305857 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs podName:6e3c0ac4-80fa-4d8d-864d-38056dc22402 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:22.805832779 +0000 UTC m=+756.114672580 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs") pod "openstack-operator-controller-manager-69bc8c84fd-8f9kj" (UID: "6e3c0ac4-80fa-4d8d-864d-38056dc22402") : secret "metrics-server-cert" not found Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.305894 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd"] Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.306599 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:22 crc kubenswrapper[4872]: E0126 09:20:22.306725 4872 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 26 09:20:22 crc kubenswrapper[4872]: E0126 09:20:22.306757 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs podName:6e3c0ac4-80fa-4d8d-864d-38056dc22402 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:22.806746782 +0000 UTC m=+756.115586583 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs") pod "openstack-operator-controller-manager-69bc8c84fd-8f9kj" (UID: "6e3c0ac4-80fa-4d8d-864d-38056dc22402") : secret "webhook-server-cert" not found Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.311237 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-jsmg6" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.344712 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vgsqx\" (UniqueName: \"kubernetes.io/projected/ced935ab-f3e5-470d-aa18-e15bb944fb86-kube-api-access-vgsqx\") pod \"watcher-operator-controller-manager-564965969-hfj2p\" (UID: \"ced935ab-f3e5-470d-aa18-e15bb944fb86\") " pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.357958 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmb45\" (UniqueName: \"kubernetes.io/projected/6e3c0ac4-80fa-4d8d-864d-38056dc22402-kube-api-access-lmb45\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.427580 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv\" (UID: \"a0993a22-0883-4db0-9c07-c30657d9fb37\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.427712 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zxs6\" (UniqueName: \"kubernetes.io/projected/1fa3edb5-2354-4c32-ac85-d6be3a7dc16c-kube-api-access-5zxs6\") pod \"rabbitmq-cluster-operator-manager-668c99d594-hmfxd\" (UID: \"1fa3edb5-2354-4c32-ac85-d6be3a7dc16c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd" Jan 26 09:20:22 crc kubenswrapper[4872]: E0126 09:20:22.429024 4872 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 09:20:22 crc kubenswrapper[4872]: E0126 09:20:22.429093 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert podName:a0993a22-0883-4db0-9c07-c30657d9fb37 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:23.429073202 +0000 UTC m=+756.737913003 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" (UID: "a0993a22-0883-4db0-9c07-c30657d9fb37") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.429510 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.482200 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zxs6\" (UniqueName: \"kubernetes.io/projected/1fa3edb5-2354-4c32-ac85-d6be3a7dc16c-kube-api-access-5zxs6\") pod \"rabbitmq-cluster-operator-manager-668c99d594-hmfxd\" (UID: \"1fa3edb5-2354-4c32-ac85-d6be3a7dc16c\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.482417 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.506406 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757"] Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.511657 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh"] Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.759786 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.863734 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:22 crc kubenswrapper[4872]: I0126 09:20:22.864219 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:22 crc kubenswrapper[4872]: E0126 09:20:22.865155 4872 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 26 09:20:22 crc kubenswrapper[4872]: E0126 09:20:22.865214 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs podName:6e3c0ac4-80fa-4d8d-864d-38056dc22402 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:23.865196387 +0000 UTC m=+757.174036188 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs") pod "openstack-operator-controller-manager-69bc8c84fd-8f9kj" (UID: "6e3c0ac4-80fa-4d8d-864d-38056dc22402") : secret "metrics-server-cert" not found Jan 26 09:20:22 crc kubenswrapper[4872]: E0126 09:20:22.865267 4872 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 26 09:20:22 crc kubenswrapper[4872]: E0126 09:20:22.865288 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs podName:6e3c0ac4-80fa-4d8d-864d-38056dc22402 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:23.865281769 +0000 UTC m=+757.174121570 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs") pod "openstack-operator-controller-manager-69bc8c84fd-8f9kj" (UID: "6e3c0ac4-80fa-4d8d-864d-38056dc22402") : secret "webhook-server-cert" not found Jan 26 09:20:23 crc kubenswrapper[4872]: I0126 09:20:23.175063 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert\") pod \"infra-operator-controller-manager-758868c854-rss8g\" (UID: \"5619ad3b-5066-4163-9cff-7654030fc5b1\") " pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:23 crc kubenswrapper[4872]: E0126 09:20:23.175324 4872 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 26 09:20:23 crc kubenswrapper[4872]: E0126 09:20:23.175402 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert podName:5619ad3b-5066-4163-9cff-7654030fc5b1 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:25.17537625 +0000 UTC m=+758.484216051 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert") pod "infra-operator-controller-manager-758868c854-rss8g" (UID: "5619ad3b-5066-4163-9cff-7654030fc5b1") : secret "infra-operator-webhook-server-cert" not found Jan 26 09:20:23 crc kubenswrapper[4872]: I0126 09:20:23.487163 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv\" (UID: \"a0993a22-0883-4db0-9c07-c30657d9fb37\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:23 crc kubenswrapper[4872]: E0126 09:20:23.487601 4872 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 09:20:23 crc kubenswrapper[4872]: E0126 09:20:23.487775 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert podName:a0993a22-0883-4db0-9c07-c30657d9fb37 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:25.48775551 +0000 UTC m=+758.796595311 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" (UID: "a0993a22-0883-4db0-9c07-c30657d9fb37") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 09:20:23 crc kubenswrapper[4872]: I0126 09:20:23.551260 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh" event={"ID":"a313f841-7d48-4621-a385-b001d8cfa11b","Type":"ContainerStarted","Data":"eaea27944ebd871305a08c49ddcd6b6c096bc97921e257f486e27fb5d43d08ff"} Jan 26 09:20:23 crc kubenswrapper[4872]: I0126 09:20:23.553493 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757" event={"ID":"866888fd-6bb1-4184-9e17-352cb8d3727f","Type":"ContainerStarted","Data":"7ec708b01cf658dd9d6198709f9e8bd72069f128c88f402d73538d688052bf11"} Jan 26 09:20:23 crc kubenswrapper[4872]: I0126 09:20:23.763953 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m"] Jan 26 09:20:23 crc kubenswrapper[4872]: W0126 09:20:23.792481 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bda61c9_b3dd_476a_86fb_882a5f172ba9.slice/crio-1b97344e151acda158c4dace6cecab50d29ab8d1c6f090bd0b15f551d7f1d514 WatchSource:0}: Error finding container 1b97344e151acda158c4dace6cecab50d29ab8d1c6f090bd0b15f551d7f1d514: Status 404 returned error can't find the container with id 1b97344e151acda158c4dace6cecab50d29ab8d1c6f090bd0b15f551d7f1d514 Jan 26 09:20:23 crc kubenswrapper[4872]: I0126 09:20:23.895760 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:23 crc kubenswrapper[4872]: I0126 09:20:23.895945 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:23 crc kubenswrapper[4872]: E0126 09:20:23.896020 4872 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 26 09:20:23 crc kubenswrapper[4872]: E0126 09:20:23.896112 4872 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 26 09:20:23 crc kubenswrapper[4872]: E0126 09:20:23.896121 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs podName:6e3c0ac4-80fa-4d8d-864d-38056dc22402 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:25.896095793 +0000 UTC m=+759.204935594 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs") pod "openstack-operator-controller-manager-69bc8c84fd-8f9kj" (UID: "6e3c0ac4-80fa-4d8d-864d-38056dc22402") : secret "webhook-server-cert" not found Jan 26 09:20:23 crc kubenswrapper[4872]: E0126 09:20:23.896174 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs podName:6e3c0ac4-80fa-4d8d-864d-38056dc22402 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:25.896163925 +0000 UTC m=+759.205003726 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs") pod "openstack-operator-controller-manager-69bc8c84fd-8f9kj" (UID: "6e3c0ac4-80fa-4d8d-864d-38056dc22402") : secret "metrics-server-cert" not found Jan 26 09:20:23 crc kubenswrapper[4872]: I0126 09:20:23.945165 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb"] Jan 26 09:20:23 crc kubenswrapper[4872]: W0126 09:20:23.952017 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f35eab0_a711_4b5d_8b5a_05c72414c30c.slice/crio-349ee9ff281e04c490d02625498fe131bd9f96367a4c890af0e3cbdefcd8accc WatchSource:0}: Error finding container 349ee9ff281e04c490d02625498fe131bd9f96367a4c890af0e3cbdefcd8accc: Status 404 returned error can't find the container with id 349ee9ff281e04c490d02625498fe131bd9f96367a4c890af0e3cbdefcd8accc Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.241026 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt"] Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.260493 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq"] Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.270641 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh"] Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.276214 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s"] Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.283899 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p"] Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.288727 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z"] Jan 26 09:20:24 crc kubenswrapper[4872]: W0126 09:20:24.291628 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3751f8d_5921_4b5f_b740_1ba5fc532f7b.slice/crio-23763dfa1f67177c5b1ccc47ff330694c5c4e50de16c2c599cbeaf278d8ff7ea WatchSource:0}: Error finding container 23763dfa1f67177c5b1ccc47ff330694c5c4e50de16c2c599cbeaf278d8ff7ea: Status 404 returned error can't find the container with id 23763dfa1f67177c5b1ccc47ff330694c5c4e50de16c2c599cbeaf278d8ff7ea Jan 26 09:20:24 crc kubenswrapper[4872]: W0126 09:20:24.298093 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60fd216c_7ecc_477f_ac29_b7334ea5e450.slice/crio-c0fb94e9fb063542a53b9e31c3eb57675e5b085a96e5f86d6c5a8efe0e1521c9 WatchSource:0}: Error finding container c0fb94e9fb063542a53b9e31c3eb57675e5b085a96e5f86d6c5a8efe0e1521c9: Status 404 returned error can't find the container with id c0fb94e9fb063542a53b9e31c3eb57675e5b085a96e5f86d6c5a8efe0e1521c9 Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.310722 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49"] Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.352444 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2"] Jan 26 09:20:24 crc kubenswrapper[4872]: W0126 09:20:24.380232 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40ff0d90_ae83_47b6_b6b0_8969b47bcf92.slice/crio-359e302769c3c17f16f97f04cb5a3d0d720a9a760d6278ad352df22428fc5158 WatchSource:0}: Error finding container 359e302769c3c17f16f97f04cb5a3d0d720a9a760d6278ad352df22428fc5158: Status 404 returned error can't find the container with id 359e302769c3c17f16f97f04cb5a3d0d720a9a760d6278ad352df22428fc5158 Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.382087 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns"] Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.398270 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g"] Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.411282 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:8abfbec47f0119a6c22c61a0ff80a4b1c6c14439a327bc75d4c529c5d8f59658,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kqj9c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-7bdb645866-vz57m_openstack-operators(acf08f03-b388-4e2a-99a2-2d492e490e67): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.411452 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ws687,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-547cbdb99f-cqr4g_openstack-operators(d92c1258-d114-4696-a481-f9fff7faa046): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.412709 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" podUID="acf08f03-b388-4e2a-99a2-2d492e490e67" Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.412762 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" podUID="d92c1258-d114-4696-a481-f9fff7faa046" Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.413337 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:b673f00227298dcfa89abb46f8296a0825add42da41e8a4bf4dd13367c738d84,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-n44p8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-6b9fb5fdcb-tcc58_openstack-operators(40ff0d90-ae83-47b6-b6b0-8969b47bcf92): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.414185 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:e02722d7581bfe1c5fc13e2fa6811d8665102ba86635c77547abf6b933cde127,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x7pl7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-85cd9769bb-zfvrw_openstack-operators(4b97d503-908a-48b0-bb0c-5e2cdcf251af): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.414568 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58"] Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.414751 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" podUID="40ff0d90-ae83-47b6-b6b0-8969b47bcf92" Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.415135 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:7869203f6f97de780368d507636031090fed3b658d2f7771acbd4481bdfc870b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vgsqx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-564965969-hfj2p_openstack-operators(ced935ab-f3e5-470d-aa18-e15bb944fb86): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.415348 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" podUID="4b97d503-908a-48b0-bb0c-5e2cdcf251af" Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.416850 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" podUID="ced935ab-f3e5-470d-aa18-e15bb944fb86" Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.419596 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5zxs6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-hmfxd_openstack-operators(1fa3edb5-2354-4c32-ac85-d6be3a7dc16c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.420813 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd" podUID="1fa3edb5-2354-4c32-ac85-d6be3a7dc16c" Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.423231 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd"] Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.435154 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-564965969-hfj2p"] Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.456201 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2"] Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.469397 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw"] Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.474007 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m"] Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.567625 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb" event={"ID":"7f35eab0-a711-4b5d-8b5a-05c72414c30c","Type":"ContainerStarted","Data":"349ee9ff281e04c490d02625498fe131bd9f96367a4c890af0e3cbdefcd8accc"} Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.569934 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq" event={"ID":"b3751f8d-5921-4b5f-b740-1ba5fc532f7b","Type":"ContainerStarted","Data":"23763dfa1f67177c5b1ccc47ff330694c5c4e50de16c2c599cbeaf278d8ff7ea"} Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.571867 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh" event={"ID":"c5bfadfb-1ac5-475e-8b5f-8096dd00d850","Type":"ContainerStarted","Data":"36fcb1ec7befa8486f9f714f81d7d70fd9e512619b5cc76a13c874f876ff85ec"} Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.573696 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt" event={"ID":"87e00bbc-66f4-4921-94ce-24d7b658341c","Type":"ContainerStarted","Data":"845955d4f43a0cc9b772e69a6dcc8837b81407669ad0e399d7f38ffa8f4dd9e2"} Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.576429 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49" event={"ID":"60fd216c-7ecc-477f-ac29-b7334ea5e450","Type":"ContainerStarted","Data":"c0fb94e9fb063542a53b9e31c3eb57675e5b085a96e5f86d6c5a8efe0e1521c9"} Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.583390 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p" event={"ID":"148da344-f5f5-43ab-ba3c-342f6cc10b47","Type":"ContainerStarted","Data":"f4f198ee70375eb0f0f075fc921d319cf7ffb58906deeb551c8a7df547352c4f"} Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.590063 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" event={"ID":"40ff0d90-ae83-47b6-b6b0-8969b47bcf92","Type":"ContainerStarted","Data":"359e302769c3c17f16f97f04cb5a3d0d720a9a760d6278ad352df22428fc5158"} Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.591714 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m" event={"ID":"2bda61c9-b3dd-476a-86fb-882a5f172ba9","Type":"ContainerStarted","Data":"1b97344e151acda158c4dace6cecab50d29ab8d1c6f090bd0b15f551d7f1d514"} Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.592005 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:b673f00227298dcfa89abb46f8296a0825add42da41e8a4bf4dd13367c738d84\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" podUID="40ff0d90-ae83-47b6-b6b0-8969b47bcf92" Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.594465 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" event={"ID":"ced935ab-f3e5-470d-aa18-e15bb944fb86","Type":"ContainerStarted","Data":"af487864877b83744ea363d131cbc8ee297d38a9e919ea942ce67a16a57c9be9"} Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.596353 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7869203f6f97de780368d507636031090fed3b658d2f7771acbd4481bdfc870b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" podUID="ced935ab-f3e5-470d-aa18-e15bb944fb86" Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.596539 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2" event={"ID":"9cdfe7c6-5b86-4204-b9bf-3597361af9fe","Type":"ContainerStarted","Data":"6efc80f7e6494dfedf3dcabd2a75f67f64b3feb05af3a7c218eb4ed4e7e01742"} Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.598348 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2" event={"ID":"37007861-ed2f-4453-8a25-aa39f6b0f347","Type":"ContainerStarted","Data":"c37e4f55e27e6a201e7b3a05bdda3a55b1bbb8733b046c0faa4a760111b53eb4"} Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.599869 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns" event={"ID":"12d27df7-ee8b-4e48-b0fe-0b1f018e71f8","Type":"ContainerStarted","Data":"2414bc84ec04a9aab120f56e6e19c0fe0d0462c7b58c686079433e15f0a2369b"} Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.602642 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd" event={"ID":"1fa3edb5-2354-4c32-ac85-d6be3a7dc16c","Type":"ContainerStarted","Data":"cc3d1ef2d416b2eb28c3fcf788c5cc0786c1e0c93e580071670a2d2e2f95c77d"} Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.604644 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd" podUID="1fa3edb5-2354-4c32-ac85-d6be3a7dc16c" Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.606885 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" event={"ID":"acf08f03-b388-4e2a-99a2-2d492e490e67","Type":"ContainerStarted","Data":"ee3a30c7906c9ddccd4076fce1efa9ccf08cea0fc037c9e981805ebab79ad5f8"} Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.610191 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:8abfbec47f0119a6c22c61a0ff80a4b1c6c14439a327bc75d4c529c5d8f59658\\\"\"" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" podUID="acf08f03-b388-4e2a-99a2-2d492e490e67" Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.615579 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s" event={"ID":"4d3db7ee-d472-427c-96a2-1c0f67a3b507","Type":"ContainerStarted","Data":"a3037e30f7e493b39ad98b4cbd93de1000942c8d708734f4b6ece4e0e9a5cbf4"} Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.618122 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z" event={"ID":"c05670e5-8312-46ce-9fe7-da7ecc691465","Type":"ContainerStarted","Data":"33208415d26a4edd306f5c79a6fbe7dee6a93f195ada0d3db9799cb9e9017ab1"} Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.621777 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" event={"ID":"d92c1258-d114-4696-a481-f9fff7faa046","Type":"ContainerStarted","Data":"a4be1acfea740f95bd7b71c418012c1a7e623e88310e5584899a60337b43d6b8"} Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.623257 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922\\\"\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" podUID="d92c1258-d114-4696-a481-f9fff7faa046" Jan 26 09:20:24 crc kubenswrapper[4872]: I0126 09:20:24.626992 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" event={"ID":"4b97d503-908a-48b0-bb0c-5e2cdcf251af","Type":"ContainerStarted","Data":"ec5ea233e6c5c9990c0567bf6385e7550963a9021d94c32f0aa63a4e2d0289e9"} Jan 26 09:20:24 crc kubenswrapper[4872]: E0126 09:20:24.631315 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:e02722d7581bfe1c5fc13e2fa6811d8665102ba86635c77547abf6b933cde127\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" podUID="4b97d503-908a-48b0-bb0c-5e2cdcf251af" Jan 26 09:20:25 crc kubenswrapper[4872]: I0126 09:20:25.219066 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert\") pod \"infra-operator-controller-manager-758868c854-rss8g\" (UID: \"5619ad3b-5066-4163-9cff-7654030fc5b1\") " pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.219369 4872 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.219443 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert podName:5619ad3b-5066-4163-9cff-7654030fc5b1 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:29.219428105 +0000 UTC m=+762.528267906 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert") pod "infra-operator-controller-manager-758868c854-rss8g" (UID: "5619ad3b-5066-4163-9cff-7654030fc5b1") : secret "infra-operator-webhook-server-cert" not found Jan 26 09:20:25 crc kubenswrapper[4872]: I0126 09:20:25.526256 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv\" (UID: \"a0993a22-0883-4db0-9c07-c30657d9fb37\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.527474 4872 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.527715 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert podName:a0993a22-0883-4db0-9c07-c30657d9fb37 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:29.52768987 +0000 UTC m=+762.836529671 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" (UID: "a0993a22-0883-4db0-9c07-c30657d9fb37") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.645575 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:445e951df2f21df6d33a466f75917e0f6103052ae751ae11887136e8ab165922\\\"\"" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" podUID="d92c1258-d114-4696-a481-f9fff7faa046" Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.646156 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:8abfbec47f0119a6c22c61a0ff80a4b1c6c14439a327bc75d4c529c5d8f59658\\\"\"" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" podUID="acf08f03-b388-4e2a-99a2-2d492e490e67" Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.646164 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd" podUID="1fa3edb5-2354-4c32-ac85-d6be3a7dc16c" Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.646239 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:b673f00227298dcfa89abb46f8296a0825add42da41e8a4bf4dd13367c738d84\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" podUID="40ff0d90-ae83-47b6-b6b0-8969b47bcf92" Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.646291 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:7869203f6f97de780368d507636031090fed3b658d2f7771acbd4481bdfc870b\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" podUID="ced935ab-f3e5-470d-aa18-e15bb944fb86" Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.648305 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:e02722d7581bfe1c5fc13e2fa6811d8665102ba86635c77547abf6b933cde127\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" podUID="4b97d503-908a-48b0-bb0c-5e2cdcf251af" Jan 26 09:20:25 crc kubenswrapper[4872]: I0126 09:20:25.938306 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:25 crc kubenswrapper[4872]: I0126 09:20:25.938448 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.938601 4872 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.938713 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs podName:6e3c0ac4-80fa-4d8d-864d-38056dc22402 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:29.938689901 +0000 UTC m=+763.247529782 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs") pod "openstack-operator-controller-manager-69bc8c84fd-8f9kj" (UID: "6e3c0ac4-80fa-4d8d-864d-38056dc22402") : secret "metrics-server-cert" not found Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.938616 4872 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 26 09:20:25 crc kubenswrapper[4872]: E0126 09:20:25.938835 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs podName:6e3c0ac4-80fa-4d8d-864d-38056dc22402 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:29.938780873 +0000 UTC m=+763.247620674 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs") pod "openstack-operator-controller-manager-69bc8c84fd-8f9kj" (UID: "6e3c0ac4-80fa-4d8d-864d-38056dc22402") : secret "webhook-server-cert" not found Jan 26 09:20:29 crc kubenswrapper[4872]: I0126 09:20:29.225276 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert\") pod \"infra-operator-controller-manager-758868c854-rss8g\" (UID: \"5619ad3b-5066-4163-9cff-7654030fc5b1\") " pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:29 crc kubenswrapper[4872]: E0126 09:20:29.225746 4872 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Jan 26 09:20:29 crc kubenswrapper[4872]: E0126 09:20:29.226438 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert podName:5619ad3b-5066-4163-9cff-7654030fc5b1 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:37.226398595 +0000 UTC m=+770.535238546 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert") pod "infra-operator-controller-manager-758868c854-rss8g" (UID: "5619ad3b-5066-4163-9cff-7654030fc5b1") : secret "infra-operator-webhook-server-cert" not found Jan 26 09:20:29 crc kubenswrapper[4872]: I0126 09:20:29.530487 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv\" (UID: \"a0993a22-0883-4db0-9c07-c30657d9fb37\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:29 crc kubenswrapper[4872]: E0126 09:20:29.530714 4872 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 09:20:29 crc kubenswrapper[4872]: E0126 09:20:29.530853 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert podName:a0993a22-0883-4db0-9c07-c30657d9fb37 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:37.530817113 +0000 UTC m=+770.839656974 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert") pod "openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" (UID: "a0993a22-0883-4db0-9c07-c30657d9fb37") : secret "openstack-baremetal-operator-webhook-server-cert" not found Jan 26 09:20:30 crc kubenswrapper[4872]: I0126 09:20:30.039162 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:30 crc kubenswrapper[4872]: I0126 09:20:30.039315 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:30 crc kubenswrapper[4872]: E0126 09:20:30.039500 4872 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Jan 26 09:20:30 crc kubenswrapper[4872]: E0126 09:20:30.039567 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs podName:6e3c0ac4-80fa-4d8d-864d-38056dc22402 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:38.039547412 +0000 UTC m=+771.348387213 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs") pod "openstack-operator-controller-manager-69bc8c84fd-8f9kj" (UID: "6e3c0ac4-80fa-4d8d-864d-38056dc22402") : secret "metrics-server-cert" not found Jan 26 09:20:30 crc kubenswrapper[4872]: E0126 09:20:30.040040 4872 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 26 09:20:30 crc kubenswrapper[4872]: E0126 09:20:30.040083 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs podName:6e3c0ac4-80fa-4d8d-864d-38056dc22402 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:38.040070215 +0000 UTC m=+771.348910016 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs") pod "openstack-operator-controller-manager-69bc8c84fd-8f9kj" (UID: "6e3c0ac4-80fa-4d8d-864d-38056dc22402") : secret "webhook-server-cert" not found Jan 26 09:20:36 crc kubenswrapper[4872]: E0126 09:20:36.031197 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.136:5001/openstack-k8s-operators/ironic-operator:5a447c5cb26272ff66e7396744825243ba48da79" Jan 26 09:20:36 crc kubenswrapper[4872]: E0126 09:20:36.031914 4872 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.136:5001/openstack-k8s-operators/ironic-operator:5a447c5cb26272ff66e7396744825243ba48da79" Jan 26 09:20:36 crc kubenswrapper[4872]: E0126 09:20:36.032154 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.136:5001/openstack-k8s-operators/ironic-operator:5a447c5cb26272ff66e7396744825243ba48da79,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-km4cd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6b46765b57-crcq2_openstack-operators(9cdfe7c6-5b86-4204-b9bf-3597361af9fe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:20:36 crc kubenswrapper[4872]: E0126 09:20:36.034035 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2" podUID="9cdfe7c6-5b86-4204-b9bf-3597361af9fe" Jan 26 09:20:36 crc kubenswrapper[4872]: E0126 09:20:36.933088 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:9caae9b3ee328df678baa26454e45e47693acdadb27f9c635680597aaec43337" Jan 26 09:20:36 crc kubenswrapper[4872]: E0126 09:20:36.933346 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:9caae9b3ee328df678baa26454e45e47693acdadb27f9c635680597aaec43337,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lv9nh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-78fdd796fd-cwkcb_openstack-operators(7f35eab0-a711-4b5d-8b5a-05c72414c30c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:20:36 crc kubenswrapper[4872]: E0126 09:20:36.934596 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb" podUID="7f35eab0-a711-4b5d-8b5a-05c72414c30c" Jan 26 09:20:36 crc kubenswrapper[4872]: E0126 09:20:36.986050 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:9caae9b3ee328df678baa26454e45e47693acdadb27f9c635680597aaec43337\\\"\"" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb" podUID="7f35eab0-a711-4b5d-8b5a-05c72414c30c" Jan 26 09:20:36 crc kubenswrapper[4872]: E0126 09:20:36.986167 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.136:5001/openstack-k8s-operators/ironic-operator:5a447c5cb26272ff66e7396744825243ba48da79\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2" podUID="9cdfe7c6-5b86-4204-b9bf-3597361af9fe" Jan 26 09:20:37 crc kubenswrapper[4872]: I0126 09:20:37.243556 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert\") pod \"infra-operator-controller-manager-758868c854-rss8g\" (UID: \"5619ad3b-5066-4163-9cff-7654030fc5b1\") " pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:37 crc kubenswrapper[4872]: I0126 09:20:37.255012 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5619ad3b-5066-4163-9cff-7654030fc5b1-cert\") pod \"infra-operator-controller-manager-758868c854-rss8g\" (UID: \"5619ad3b-5066-4163-9cff-7654030fc5b1\") " pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:37 crc kubenswrapper[4872]: I0126 09:20:37.282325 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:37 crc kubenswrapper[4872]: I0126 09:20:37.550347 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv\" (UID: \"a0993a22-0883-4db0-9c07-c30657d9fb37\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:37 crc kubenswrapper[4872]: I0126 09:20:37.555350 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a0993a22-0883-4db0-9c07-c30657d9fb37-cert\") pod \"openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv\" (UID: \"a0993a22-0883-4db0-9c07-c30657d9fb37\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:37 crc kubenswrapper[4872]: I0126 09:20:37.649023 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:38 crc kubenswrapper[4872]: I0126 09:20:38.057523 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:38 crc kubenswrapper[4872]: I0126 09:20:38.057625 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:38 crc kubenswrapper[4872]: E0126 09:20:38.057832 4872 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Jan 26 09:20:38 crc kubenswrapper[4872]: E0126 09:20:38.057897 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs podName:6e3c0ac4-80fa-4d8d-864d-38056dc22402 nodeName:}" failed. No retries permitted until 2026-01-26 09:20:54.057877052 +0000 UTC m=+787.366716853 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs") pod "openstack-operator-controller-manager-69bc8c84fd-8f9kj" (UID: "6e3c0ac4-80fa-4d8d-864d-38056dc22402") : secret "webhook-server-cert" not found Jan 26 09:20:38 crc kubenswrapper[4872]: I0126 09:20:38.073764 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-metrics-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:40 crc kubenswrapper[4872]: I0126 09:20:40.227698 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:20:40 crc kubenswrapper[4872]: I0126 09:20:40.227779 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:20:40 crc kubenswrapper[4872]: I0126 09:20:40.227892 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:20:40 crc kubenswrapper[4872]: I0126 09:20:40.228700 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df957db96f1d4f96c1bc63715f9b57ffc3271e0d6dac35d880b93d7745dd43b0"} pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 09:20:40 crc kubenswrapper[4872]: I0126 09:20:40.228762 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" containerID="cri-o://df957db96f1d4f96c1bc63715f9b57ffc3271e0d6dac35d880b93d7745dd43b0" gracePeriod=600 Jan 26 09:20:42 crc kubenswrapper[4872]: I0126 09:20:42.030615 4872 generic.go:334] "Generic (PLEG): container finished" podID="fca44d96-a000-4bf2-8283-a937b0192880" containerID="df957db96f1d4f96c1bc63715f9b57ffc3271e0d6dac35d880b93d7745dd43b0" exitCode=0 Jan 26 09:20:42 crc kubenswrapper[4872]: I0126 09:20:42.030664 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerDied","Data":"df957db96f1d4f96c1bc63715f9b57ffc3271e0d6dac35d880b93d7745dd43b0"} Jan 26 09:20:42 crc kubenswrapper[4872]: I0126 09:20:42.031152 4872 scope.go:117] "RemoveContainer" containerID="b2ddad1c18cf250920b95ba098a58a140582a8cf9f6b8c8479e40b6449d7e085" Jan 26 09:20:43 crc kubenswrapper[4872]: E0126 09:20:43.985701 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:8bee4480babd6fd8f686e0ba52a304acb6ffb90f09c7c57e7f5df5f7658836d8" Jan 26 09:20:43 crc kubenswrapper[4872]: E0126 09:20:43.985981 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:8bee4480babd6fd8f686e0ba52a304acb6ffb90f09c7c57e7f5df5f7658836d8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gvb78,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-78c6999f6f-wct5s_openstack-operators(4d3db7ee-d472-427c-96a2-1c0f67a3b507): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:20:43 crc kubenswrapper[4872]: E0126 09:20:43.987211 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s" podUID="4d3db7ee-d472-427c-96a2-1c0f67a3b507" Jan 26 09:20:44 crc kubenswrapper[4872]: E0126 09:20:44.065899 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:8bee4480babd6fd8f686e0ba52a304acb6ffb90f09c7c57e7f5df5f7658836d8\\\"\"" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s" podUID="4d3db7ee-d472-427c-96a2-1c0f67a3b507" Jan 26 09:20:45 crc kubenswrapper[4872]: E0126 09:20:45.136525 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:013c0ad82d21a21c7eece5cd4b5d5c4b8eb410b6671ac33a6f3fb78c8510811d" Jan 26 09:20:45 crc kubenswrapper[4872]: E0126 09:20:45.137338 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:013c0ad82d21a21c7eece5cd4b5d5c4b8eb410b6671ac33a6f3fb78c8510811d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s5qsw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-79d5ccc684-7fm49_openstack-operators(60fd216c-7ecc-477f-ac29-b7334ea5e450): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:20:45 crc kubenswrapper[4872]: E0126 09:20:45.139472 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49" podUID="60fd216c-7ecc-477f-ac29-b7334ea5e450" Jan 26 09:20:46 crc kubenswrapper[4872]: E0126 09:20:46.092665 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:013c0ad82d21a21c7eece5cd4b5d5c4b8eb410b6671ac33a6f3fb78c8510811d\\\"\"" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49" podUID="60fd216c-7ecc-477f-ac29-b7334ea5e450" Jan 26 09:20:46 crc kubenswrapper[4872]: E0126 09:20:46.514401 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:2f9a2f064448faebbae58f52d564dc0e8e39bed0fc12bd6b9fe925e42f1b5492" Jan 26 09:20:46 crc kubenswrapper[4872]: E0126 09:20:46.515090 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:2f9a2f064448faebbae58f52d564dc0e8e39bed0fc12bd6b9fe925e42f1b5492,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pln5b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-594c8c9d5d-g5txh_openstack-operators(c5bfadfb-1ac5-475e-8b5f-8096dd00d850): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:20:46 crc kubenswrapper[4872]: E0126 09:20:46.516280 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh" podUID="c5bfadfb-1ac5-475e-8b5f-8096dd00d850" Jan 26 09:20:47 crc kubenswrapper[4872]: E0126 09:20:47.087562 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:2f9a2f064448faebbae58f52d564dc0e8e39bed0fc12bd6b9fe925e42f1b5492\\\"\"" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh" podUID="c5bfadfb-1ac5-475e-8b5f-8096dd00d850" Jan 26 09:20:47 crc kubenswrapper[4872]: E0126 09:20:47.181178 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:ed489f21a0c72557d2da5a271808f19b7c7b85ef32fd9f4aa91bdbfc5bca3bdd" Jan 26 09:20:47 crc kubenswrapper[4872]: E0126 09:20:47.181567 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:ed489f21a0c72557d2da5a271808f19b7c7b85ef32fd9f4aa91bdbfc5bca3bdd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2vf26,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-5f4cd88d46-77t7p_openstack-operators(148da344-f5f5-43ab-ba3c-342f6cc10b47): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:20:47 crc kubenswrapper[4872]: E0126 09:20:47.183566 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p" podUID="148da344-f5f5-43ab-ba3c-342f6cc10b47" Jan 26 09:20:47 crc kubenswrapper[4872]: E0126 09:20:47.812607 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349" Jan 26 09:20:47 crc kubenswrapper[4872]: E0126 09:20:47.816149 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bbglf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-b8b6d4659-q4kjt_openstack-operators(87e00bbc-66f4-4921-94ce-24d7b658341c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:20:47 crc kubenswrapper[4872]: E0126 09:20:47.817702 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt" podUID="87e00bbc-66f4-4921-94ce-24d7b658341c" Jan 26 09:20:48 crc kubenswrapper[4872]: E0126 09:20:48.093087 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:8e340ff11922b38e811261de96982e1aff5f4eb8f225d1d9f5973025a4fe8349\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt" podUID="87e00bbc-66f4-4921-94ce-24d7b658341c" Jan 26 09:20:48 crc kubenswrapper[4872]: E0126 09:20:48.093426 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:ed489f21a0c72557d2da5a271808f19b7c7b85ef32fd9f4aa91bdbfc5bca3bdd\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p" podUID="148da344-f5f5-43ab-ba3c-342f6cc10b47" Jan 26 09:20:53 crc kubenswrapper[4872]: I0126 09:20:53.779090 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv"] Jan 26 09:20:54 crc kubenswrapper[4872]: I0126 09:20:54.099410 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:54 crc kubenswrapper[4872]: I0126 09:20:54.108680 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6e3c0ac4-80fa-4d8d-864d-38056dc22402-webhook-certs\") pod \"openstack-operator-controller-manager-69bc8c84fd-8f9kj\" (UID: \"6e3c0ac4-80fa-4d8d-864d-38056dc22402\") " pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:54 crc kubenswrapper[4872]: I0126 09:20:54.149901 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" event={"ID":"a0993a22-0883-4db0-9c07-c30657d9fb37","Type":"ContainerStarted","Data":"34aa8589e9921cdc773baee9d8f76488845cf50f8a576a21e9ef58de9d2104db"} Jan 26 09:20:54 crc kubenswrapper[4872]: I0126 09:20:54.244149 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-758868c854-rss8g"] Jan 26 09:20:54 crc kubenswrapper[4872]: W0126 09:20:54.269125 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5619ad3b_5066_4163_9cff_7654030fc5b1.slice/crio-e2a67f853455faefb8dff8938220e955dc3b043322944ecbe4f8be5b87730a38 WatchSource:0}: Error finding container e2a67f853455faefb8dff8938220e955dc3b043322944ecbe4f8be5b87730a38: Status 404 returned error can't find the container with id e2a67f853455faefb8dff8938220e955dc3b043322944ecbe4f8be5b87730a38 Jan 26 09:20:54 crc kubenswrapper[4872]: I0126 09:20:54.306162 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-l2482" Jan 26 09:20:54 crc kubenswrapper[4872]: I0126 09:20:54.312326 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.040274 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj"] Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.183069 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2" event={"ID":"37007861-ed2f-4453-8a25-aa39f6b0f347","Type":"ContainerStarted","Data":"09dcce1ba0edb3aaeadaed1dcfab6f1d3ba4d99c47e5c889d691e457f6340710"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.183926 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.219111 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" event={"ID":"40ff0d90-ae83-47b6-b6b0-8969b47bcf92","Type":"ContainerStarted","Data":"839893c22d5c41db8bd740c253362901d1880e0a0d9f6268dd45df558b29deec"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.219631 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.221052 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m" event={"ID":"2bda61c9-b3dd-476a-86fb-882a5f172ba9","Type":"ContainerStarted","Data":"6c0a6d9f187824317171268a87b436d56acfd8248f4f45cf0db34d42de30eec5"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.221854 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.223757 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2" podStartSLOduration=8.561947323 podStartE2EDuration="34.223727384s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.376074416 +0000 UTC m=+757.684914217" lastFinishedPulling="2026-01-26 09:20:50.037854477 +0000 UTC m=+783.346694278" observedRunningTime="2026-01-26 09:20:55.211162129 +0000 UTC m=+788.520001930" watchObservedRunningTime="2026-01-26 09:20:55.223727384 +0000 UTC m=+788.532567185" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.250313 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" podStartSLOduration=4.630571049 podStartE2EDuration="34.250292703s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.413245924 +0000 UTC m=+757.722085725" lastFinishedPulling="2026-01-26 09:20:54.032967578 +0000 UTC m=+787.341807379" observedRunningTime="2026-01-26 09:20:55.246148949 +0000 UTC m=+788.554988750" watchObservedRunningTime="2026-01-26 09:20:55.250292703 +0000 UTC m=+788.559132494" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.250586 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" event={"ID":"6e3c0ac4-80fa-4d8d-864d-38056dc22402","Type":"ContainerStarted","Data":"1ce8871faf1547ca014236ef7a98247e85069c000061d6373de7993778153461"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.271146 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z" event={"ID":"c05670e5-8312-46ce-9fe7-da7ecc691465","Type":"ContainerStarted","Data":"40924a9dc06de09476e887807d96aceefe4c9740408ec8f8fa80230e4b3d2f17"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.272155 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.274641 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" event={"ID":"4b97d503-908a-48b0-bb0c-5e2cdcf251af","Type":"ContainerStarted","Data":"b4764cbd1cdd29a4c4813a237db9de30bc322cdff7c9c4d690d67693c26648ab"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.275607 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.288852 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m" podStartSLOduration=8.044978974 podStartE2EDuration="34.288830085s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:23.794836816 +0000 UTC m=+757.103676617" lastFinishedPulling="2026-01-26 09:20:50.038687927 +0000 UTC m=+783.347527728" observedRunningTime="2026-01-26 09:20:55.287175753 +0000 UTC m=+788.596015554" watchObservedRunningTime="2026-01-26 09:20:55.288830085 +0000 UTC m=+788.597669886" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.325293 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2" event={"ID":"9cdfe7c6-5b86-4204-b9bf-3597361af9fe","Type":"ContainerStarted","Data":"b4890e301e5e1c6bc61ae7e160b8e3903b38c1677985d6705be034091ddae2a5"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.326398 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.345479 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq" event={"ID":"b3751f8d-5921-4b5f-b740-1ba5fc532f7b","Type":"ContainerStarted","Data":"19d327d799fbf644b6361654227c69e4a20fa4fe818fd2385893107b978ce1f8"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.347000 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.355011 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" podStartSLOduration=4.772740981 podStartE2EDuration="34.35498396s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.414080126 +0000 UTC m=+757.722919927" lastFinishedPulling="2026-01-26 09:20:53.996323105 +0000 UTC m=+787.305162906" observedRunningTime="2026-01-26 09:20:55.347494192 +0000 UTC m=+788.656333993" watchObservedRunningTime="2026-01-26 09:20:55.35498396 +0000 UTC m=+788.663823751" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.375145 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns" event={"ID":"12d27df7-ee8b-4e48-b0fe-0b1f018e71f8","Type":"ContainerStarted","Data":"fa4026aa14b47821d87c9cb7a929772bbdde9f1e0757890f879fac66faeb0356"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.376038 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.378144 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z" podStartSLOduration=8.694448058999999 podStartE2EDuration="34.378123533s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.354985343 +0000 UTC m=+757.663825134" lastFinishedPulling="2026-01-26 09:20:50.038660807 +0000 UTC m=+783.347500608" observedRunningTime="2026-01-26 09:20:55.374195474 +0000 UTC m=+788.683035275" watchObservedRunningTime="2026-01-26 09:20:55.378123533 +0000 UTC m=+788.686963324" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.408017 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" event={"ID":"acf08f03-b388-4e2a-99a2-2d492e490e67","Type":"ContainerStarted","Data":"c734cb51a0fbfc1bbb8206c47b3c8a06d9ad82bd31c1b31f3d2269b05d5cd54b"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.408700 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.408719 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2" podStartSLOduration=4.746141908 podStartE2EDuration="34.408696084s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.355517406 +0000 UTC m=+757.664357207" lastFinishedPulling="2026-01-26 09:20:54.018071582 +0000 UTC m=+787.326911383" observedRunningTime="2026-01-26 09:20:55.406284752 +0000 UTC m=+788.715124553" watchObservedRunningTime="2026-01-26 09:20:55.408696084 +0000 UTC m=+788.717535885" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.427189 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq" podStartSLOduration=8.70647424 podStartE2EDuration="34.427161108s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.316538682 +0000 UTC m=+757.625378483" lastFinishedPulling="2026-01-26 09:20:50.03722555 +0000 UTC m=+783.346065351" observedRunningTime="2026-01-26 09:20:55.426219884 +0000 UTC m=+788.735059685" watchObservedRunningTime="2026-01-26 09:20:55.427161108 +0000 UTC m=+788.736000909" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.447316 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerStarted","Data":"ad422c6c34b0fdde5ba287cc03a049a9b7acc36305d83e2c495ecd1a880a0a0f"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.454204 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" event={"ID":"d92c1258-d114-4696-a481-f9fff7faa046","Type":"ContainerStarted","Data":"511d11fb887b9df53e14a0ec0fe6e1e82c7048b2196be31f6f28cb650ae826b5"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.454998 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.476700 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" event={"ID":"5619ad3b-5066-4163-9cff-7654030fc5b1","Type":"ContainerStarted","Data":"e2a67f853455faefb8dff8938220e955dc3b043322944ecbe4f8be5b87730a38"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.491670 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd" event={"ID":"1fa3edb5-2354-4c32-ac85-d6be3a7dc16c","Type":"ContainerStarted","Data":"c447f338b7d4e6073363a3f4b0a2439824af63e681097a24e82b143d689515e2"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.501000 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" podStartSLOduration=4.930974816 podStartE2EDuration="34.500979947s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.41108064 +0000 UTC m=+757.719920441" lastFinishedPulling="2026-01-26 09:20:53.981085781 +0000 UTC m=+787.289925572" observedRunningTime="2026-01-26 09:20:55.499366616 +0000 UTC m=+788.808206417" watchObservedRunningTime="2026-01-26 09:20:55.500979947 +0000 UTC m=+788.809819748" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.503044 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns" podStartSLOduration=8.842258194 podStartE2EDuration="34.503027699s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.376448425 +0000 UTC m=+757.685288226" lastFinishedPulling="2026-01-26 09:20:50.03721793 +0000 UTC m=+783.346057731" observedRunningTime="2026-01-26 09:20:55.4740869 +0000 UTC m=+788.782926701" watchObservedRunningTime="2026-01-26 09:20:55.503027699 +0000 UTC m=+788.811867500" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.507317 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757" event={"ID":"866888fd-6bb1-4184-9e17-352cb8d3727f","Type":"ContainerStarted","Data":"ab367cf4c66fa15a737acfc1b887c56616775480b840195cc3059b00f25eea57"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.507987 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.509084 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" event={"ID":"ced935ab-f3e5-470d-aa18-e15bb944fb86","Type":"ContainerStarted","Data":"180d0a3209b17e0ce49ecf09ec0b0e2543e723e596bc814142a3f3b57fd19800"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.509426 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.510408 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh" event={"ID":"a313f841-7d48-4621-a385-b001d8cfa11b","Type":"ContainerStarted","Data":"8b41112b36b95b449b3537ef29843a12a639bbd3c8a5b2603f72930287ee0aa3"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.510760 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.511563 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb" event={"ID":"7f35eab0-a711-4b5d-8b5a-05c72414c30c","Type":"ContainerStarted","Data":"f80d8d308bdc8e4ca4b44ec0ee7b30ad4031dd68d2bade644c6c091fbcfc0f72"} Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.511891 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.527054 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" podStartSLOduration=4.957315339 podStartE2EDuration="34.527033633s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.411247644 +0000 UTC m=+757.720087445" lastFinishedPulling="2026-01-26 09:20:53.980965938 +0000 UTC m=+787.289805739" observedRunningTime="2026-01-26 09:20:55.524470388 +0000 UTC m=+788.833310189" watchObservedRunningTime="2026-01-26 09:20:55.527033633 +0000 UTC m=+788.835873434" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.554386 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hmfxd" podStartSLOduration=3.869436946 podStartE2EDuration="33.554362452s" podCreationTimestamp="2026-01-26 09:20:22 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.419120732 +0000 UTC m=+757.727960533" lastFinishedPulling="2026-01-26 09:20:54.104046238 +0000 UTC m=+787.412886039" observedRunningTime="2026-01-26 09:20:55.551576512 +0000 UTC m=+788.860416313" watchObservedRunningTime="2026-01-26 09:20:55.554362452 +0000 UTC m=+788.863202243" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.635366 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh" podStartSLOduration=9.664813224 podStartE2EDuration="34.63534204s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:22.818455396 +0000 UTC m=+756.127295197" lastFinishedPulling="2026-01-26 09:20:47.788984212 +0000 UTC m=+781.097824013" observedRunningTime="2026-01-26 09:20:55.634569761 +0000 UTC m=+788.943409562" watchObservedRunningTime="2026-01-26 09:20:55.63534204 +0000 UTC m=+788.944181841" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.662278 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" podStartSLOduration=5.069340452 podStartE2EDuration="34.662244349s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.415068961 +0000 UTC m=+757.723908762" lastFinishedPulling="2026-01-26 09:20:54.007972858 +0000 UTC m=+787.316812659" observedRunningTime="2026-01-26 09:20:55.657403346 +0000 UTC m=+788.966243147" watchObservedRunningTime="2026-01-26 09:20:55.662244349 +0000 UTC m=+788.971084150" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.697534 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb" podStartSLOduration=4.644382086 podStartE2EDuration="34.697511817s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:23.954841477 +0000 UTC m=+757.263681268" lastFinishedPulling="2026-01-26 09:20:54.007971198 +0000 UTC m=+787.316810999" observedRunningTime="2026-01-26 09:20:55.694831039 +0000 UTC m=+789.003670840" watchObservedRunningTime="2026-01-26 09:20:55.697511817 +0000 UTC m=+789.006351618" Jan 26 09:20:55 crc kubenswrapper[4872]: I0126 09:20:55.734854 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757" podStartSLOduration=7.498786636 podStartE2EDuration="34.734833857s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:22.801232941 +0000 UTC m=+756.110072742" lastFinishedPulling="2026-01-26 09:20:50.037280162 +0000 UTC m=+783.346119963" observedRunningTime="2026-01-26 09:20:55.734539999 +0000 UTC m=+789.043379800" watchObservedRunningTime="2026-01-26 09:20:55.734833857 +0000 UTC m=+789.043673658" Jan 26 09:20:56 crc kubenswrapper[4872]: I0126 09:20:56.524340 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" event={"ID":"6e3c0ac4-80fa-4d8d-864d-38056dc22402","Type":"ContainerStarted","Data":"9f0a933041bb723e29e2a27bedcebcc4d80b9ade227cce45bf98526318b996eb"} Jan 26 09:20:56 crc kubenswrapper[4872]: I0126 09:20:56.554227 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" podStartSLOduration=34.554204361000004 podStartE2EDuration="34.554204361s" podCreationTimestamp="2026-01-26 09:20:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:20:56.551420811 +0000 UTC m=+789.860260612" watchObservedRunningTime="2026-01-26 09:20:56.554204361 +0000 UTC m=+789.863044162" Jan 26 09:20:57 crc kubenswrapper[4872]: I0126 09:20:57.539385 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:20:59 crc kubenswrapper[4872]: I0126 09:20:59.580348 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" event={"ID":"5619ad3b-5066-4163-9cff-7654030fc5b1","Type":"ContainerStarted","Data":"cc478bc1269463d551d2e0df44b417aabbcd436b2f6e45340058368e31e91f99"} Jan 26 09:20:59 crc kubenswrapper[4872]: I0126 09:20:59.582457 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:20:59 crc kubenswrapper[4872]: I0126 09:20:59.589337 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s" event={"ID":"4d3db7ee-d472-427c-96a2-1c0f67a3b507","Type":"ContainerStarted","Data":"74f2fcebdc3bbb3fc30dd4f3f47b3845531c6a69268a670cd69e4f67fffc0fe2"} Jan 26 09:20:59 crc kubenswrapper[4872]: I0126 09:20:59.590131 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s" Jan 26 09:20:59 crc kubenswrapper[4872]: I0126 09:20:59.591489 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" event={"ID":"a0993a22-0883-4db0-9c07-c30657d9fb37","Type":"ContainerStarted","Data":"cc262aa0db80b0819215440a06ed5daf513cac6aaecd92b1c0bcfe2b43096568"} Jan 26 09:20:59 crc kubenswrapper[4872]: I0126 09:20:59.591649 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:20:59 crc kubenswrapper[4872]: I0126 09:20:59.608269 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" podStartSLOduration=33.983601837 podStartE2EDuration="38.608242971s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:54.275025373 +0000 UTC m=+787.583865184" lastFinishedPulling="2026-01-26 09:20:58.899666517 +0000 UTC m=+792.208506318" observedRunningTime="2026-01-26 09:20:59.602140048 +0000 UTC m=+792.910979849" watchObservedRunningTime="2026-01-26 09:20:59.608242971 +0000 UTC m=+792.917082772" Jan 26 09:20:59 crc kubenswrapper[4872]: I0126 09:20:59.635012 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s" podStartSLOduration=4.039522568 podStartE2EDuration="38.634975934s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.303222446 +0000 UTC m=+757.612062247" lastFinishedPulling="2026-01-26 09:20:58.898675812 +0000 UTC m=+792.207515613" observedRunningTime="2026-01-26 09:20:59.628082161 +0000 UTC m=+792.936921962" watchObservedRunningTime="2026-01-26 09:20:59.634975934 +0000 UTC m=+792.943815735" Jan 26 09:21:00 crc kubenswrapper[4872]: I0126 09:21:00.212768 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" podStartSLOduration=34.289160882 podStartE2EDuration="39.212739525s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:53.976197787 +0000 UTC m=+787.285037588" lastFinishedPulling="2026-01-26 09:20:58.89977639 +0000 UTC m=+792.208616231" observedRunningTime="2026-01-26 09:20:59.657932942 +0000 UTC m=+792.966772763" watchObservedRunningTime="2026-01-26 09:21:00.212739525 +0000 UTC m=+793.521579336" Jan 26 09:21:00 crc kubenswrapper[4872]: I0126 09:21:00.599960 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49" event={"ID":"60fd216c-7ecc-477f-ac29-b7334ea5e450","Type":"ContainerStarted","Data":"31ba7d708cebf0e5f6b13f6369f473ca881e3703fc67dad725ffbc7a1188b9e0"} Jan 26 09:21:00 crc kubenswrapper[4872]: I0126 09:21:00.600730 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49" Jan 26 09:21:00 crc kubenswrapper[4872]: I0126 09:21:00.638309 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49" podStartSLOduration=4.328162287 podStartE2EDuration="39.638282122s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.303331768 +0000 UTC m=+757.612171569" lastFinishedPulling="2026-01-26 09:20:59.613451583 +0000 UTC m=+792.922291404" observedRunningTime="2026-01-26 09:21:00.632222519 +0000 UTC m=+793.941062320" watchObservedRunningTime="2026-01-26 09:21:00.638282122 +0000 UTC m=+793.947121933" Jan 26 09:21:01 crc kubenswrapper[4872]: I0126 09:21:01.520926 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7f86f8796f-96757" Jan 26 09:21:01 crc kubenswrapper[4872]: I0126 09:21:01.539792 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7478f7dbf9-cvcnq" Jan 26 09:21:01 crc kubenswrapper[4872]: I0126 09:21:01.552038 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-b45d7bf98-ldqqh" Jan 26 09:21:01 crc kubenswrapper[4872]: I0126 09:21:01.587213 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-78fdd796fd-cwkcb" Jan 26 09:21:01 crc kubenswrapper[4872]: I0126 09:21:01.629555 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh" event={"ID":"c5bfadfb-1ac5-475e-8b5f-8096dd00d850","Type":"ContainerStarted","Data":"5e60c51da1dbee8fc245c4b29156541055d62eb7344ad09589b150fff678ff03"} Jan 26 09:21:01 crc kubenswrapper[4872]: I0126 09:21:01.633753 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh" Jan 26 09:21:01 crc kubenswrapper[4872]: I0126 09:21:01.651864 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh" podStartSLOduration=4.345810783 podStartE2EDuration="40.651839946s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.31883361 +0000 UTC m=+757.627673411" lastFinishedPulling="2026-01-26 09:21:00.624862763 +0000 UTC m=+793.933702574" observedRunningTime="2026-01-26 09:21:01.649243481 +0000 UTC m=+794.958083292" watchObservedRunningTime="2026-01-26 09:21:01.651839946 +0000 UTC m=+794.960679747" Jan 26 09:21:01 crc kubenswrapper[4872]: I0126 09:21:01.670326 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-77d5c5b54f-s4b5m" Jan 26 09:21:01 crc kubenswrapper[4872]: I0126 09:21:01.733631 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6b46765b57-crcq2" Jan 26 09:21:02 crc kubenswrapper[4872]: I0126 09:21:02.035195 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-78d58447c5-qwp9z" Jan 26 09:21:02 crc kubenswrapper[4872]: I0126 09:21:02.035408 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-7bdb645866-vz57m" Jan 26 09:21:02 crc kubenswrapper[4872]: I0126 09:21:02.035426 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6b9fb5fdcb-tcc58" Jan 26 09:21:02 crc kubenswrapper[4872]: I0126 09:21:02.103367 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-6f75f45d54-r7rv2" Jan 26 09:21:02 crc kubenswrapper[4872]: I0126 09:21:02.246003 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-85cd9769bb-zfvrw" Jan 26 09:21:02 crc kubenswrapper[4872]: I0126 09:21:02.259263 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-69797bbcbd-tdnns" Jan 26 09:21:02 crc kubenswrapper[4872]: I0126 09:21:02.509914 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-564965969-hfj2p" Jan 26 09:21:02 crc kubenswrapper[4872]: I0126 09:21:02.510614 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-547cbdb99f-cqr4g" Jan 26 09:21:02 crc kubenswrapper[4872]: I0126 09:21:02.639838 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p" event={"ID":"148da344-f5f5-43ab-ba3c-342f6cc10b47","Type":"ContainerStarted","Data":"b24d36bd27ea27f5ddddfa5733c7a6692bec3d81d2351c6892363ec1817e53f1"} Jan 26 09:21:02 crc kubenswrapper[4872]: I0126 09:21:02.640318 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p" Jan 26 09:21:02 crc kubenswrapper[4872]: I0126 09:21:02.662753 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p" podStartSLOduration=4.374023206 podStartE2EDuration="41.662724314s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.362038691 +0000 UTC m=+757.670878492" lastFinishedPulling="2026-01-26 09:21:01.650739799 +0000 UTC m=+794.959579600" observedRunningTime="2026-01-26 09:21:02.656298571 +0000 UTC m=+795.965138372" watchObservedRunningTime="2026-01-26 09:21:02.662724314 +0000 UTC m=+795.971564115" Jan 26 09:21:03 crc kubenswrapper[4872]: I0126 09:21:03.648644 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt" event={"ID":"87e00bbc-66f4-4921-94ce-24d7b658341c","Type":"ContainerStarted","Data":"f04ef655936f7a4438a1857f8837c2aea572d44f107fe8e082620283354b784c"} Jan 26 09:21:03 crc kubenswrapper[4872]: I0126 09:21:03.649572 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt" Jan 26 09:21:04 crc kubenswrapper[4872]: I0126 09:21:04.342255 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-69bc8c84fd-8f9kj" Jan 26 09:21:04 crc kubenswrapper[4872]: I0126 09:21:04.367914 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt" podStartSLOduration=4.930613181 podStartE2EDuration="43.367888874s" podCreationTimestamp="2026-01-26 09:20:21 +0000 UTC" firstStartedPulling="2026-01-26 09:20:24.362729068 +0000 UTC m=+757.671568869" lastFinishedPulling="2026-01-26 09:21:02.800004741 +0000 UTC m=+796.108844562" observedRunningTime="2026-01-26 09:21:03.668757008 +0000 UTC m=+796.977596809" watchObservedRunningTime="2026-01-26 09:21:04.367888874 +0000 UTC m=+797.676728675" Jan 26 09:21:07 crc kubenswrapper[4872]: I0126 09:21:07.291302 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-758868c854-rss8g" Jan 26 09:21:07 crc kubenswrapper[4872]: I0126 09:21:07.654656 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv" Jan 26 09:21:11 crc kubenswrapper[4872]: I0126 09:21:11.649370 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-594c8c9d5d-g5txh" Jan 26 09:21:11 crc kubenswrapper[4872]: I0126 09:21:11.766580 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b8b6d4659-q4kjt" Jan 26 09:21:11 crc kubenswrapper[4872]: I0126 09:21:11.948921 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-78c6999f6f-wct5s" Jan 26 09:21:12 crc kubenswrapper[4872]: I0126 09:21:12.061486 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5f4cd88d46-77t7p" Jan 26 09:21:12 crc kubenswrapper[4872]: I0126 09:21:12.106464 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-79d5ccc684-7fm49" Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.869448 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-q4wxs"] Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.879300 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.887940 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-q4wxs"] Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.888406 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.888471 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-ktc94" Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.888717 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.892684 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.947050 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mfc4j"] Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.953586 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.956510 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.968136 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mfc4j"] Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.972964 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km4x2\" (UniqueName: \"kubernetes.io/projected/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23-kube-api-access-km4x2\") pod \"dnsmasq-dns-675f4bcbfc-q4wxs\" (UID: \"26bf0d4f-2ed3-4428-8a62-4b3a1947fa23\") " pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" Jan 26 09:21:29 crc kubenswrapper[4872]: I0126 09:21:29.973099 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23-config\") pod \"dnsmasq-dns-675f4bcbfc-q4wxs\" (UID: \"26bf0d4f-2ed3-4428-8a62-4b3a1947fa23\") " pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.074351 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a280b74a-cdbf-430b-9686-3055aaebf27e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-mfc4j\" (UID: \"a280b74a-cdbf-430b-9686-3055aaebf27e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.074425 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zstx7\" (UniqueName: \"kubernetes.io/projected/a280b74a-cdbf-430b-9686-3055aaebf27e-kube-api-access-zstx7\") pod \"dnsmasq-dns-78dd6ddcc-mfc4j\" (UID: \"a280b74a-cdbf-430b-9686-3055aaebf27e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.074479 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23-config\") pod \"dnsmasq-dns-675f4bcbfc-q4wxs\" (UID: \"26bf0d4f-2ed3-4428-8a62-4b3a1947fa23\") " pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.074537 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a280b74a-cdbf-430b-9686-3055aaebf27e-config\") pod \"dnsmasq-dns-78dd6ddcc-mfc4j\" (UID: \"a280b74a-cdbf-430b-9686-3055aaebf27e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.074581 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km4x2\" (UniqueName: \"kubernetes.io/projected/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23-kube-api-access-km4x2\") pod \"dnsmasq-dns-675f4bcbfc-q4wxs\" (UID: \"26bf0d4f-2ed3-4428-8a62-4b3a1947fa23\") " pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.075422 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23-config\") pod \"dnsmasq-dns-675f4bcbfc-q4wxs\" (UID: \"26bf0d4f-2ed3-4428-8a62-4b3a1947fa23\") " pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.109428 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km4x2\" (UniqueName: \"kubernetes.io/projected/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23-kube-api-access-km4x2\") pod \"dnsmasq-dns-675f4bcbfc-q4wxs\" (UID: \"26bf0d4f-2ed3-4428-8a62-4b3a1947fa23\") " pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.176081 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a280b74a-cdbf-430b-9686-3055aaebf27e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-mfc4j\" (UID: \"a280b74a-cdbf-430b-9686-3055aaebf27e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.176134 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zstx7\" (UniqueName: \"kubernetes.io/projected/a280b74a-cdbf-430b-9686-3055aaebf27e-kube-api-access-zstx7\") pod \"dnsmasq-dns-78dd6ddcc-mfc4j\" (UID: \"a280b74a-cdbf-430b-9686-3055aaebf27e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.176190 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a280b74a-cdbf-430b-9686-3055aaebf27e-config\") pod \"dnsmasq-dns-78dd6ddcc-mfc4j\" (UID: \"a280b74a-cdbf-430b-9686-3055aaebf27e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.177246 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a280b74a-cdbf-430b-9686-3055aaebf27e-config\") pod \"dnsmasq-dns-78dd6ddcc-mfc4j\" (UID: \"a280b74a-cdbf-430b-9686-3055aaebf27e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.177417 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a280b74a-cdbf-430b-9686-3055aaebf27e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-mfc4j\" (UID: \"a280b74a-cdbf-430b-9686-3055aaebf27e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.194469 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zstx7\" (UniqueName: \"kubernetes.io/projected/a280b74a-cdbf-430b-9686-3055aaebf27e-kube-api-access-zstx7\") pod \"dnsmasq-dns-78dd6ddcc-mfc4j\" (UID: \"a280b74a-cdbf-430b-9686-3055aaebf27e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.199378 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.272052 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.547975 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-q4wxs"] Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.559276 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.613221 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mfc4j"] Jan 26 09:21:30 crc kubenswrapper[4872]: W0126 09:21:30.617750 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda280b74a_cdbf_430b_9686_3055aaebf27e.slice/crio-3554a7c53fc95890d863673f987d4669b5b5018a8f17ed026dcde0978ac63eae WatchSource:0}: Error finding container 3554a7c53fc95890d863673f987d4669b5b5018a8f17ed026dcde0978ac63eae: Status 404 returned error can't find the container with id 3554a7c53fc95890d863673f987d4669b5b5018a8f17ed026dcde0978ac63eae Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.843195 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7whlt"] Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.845139 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.856434 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7whlt"] Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.910051 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" event={"ID":"26bf0d4f-2ed3-4428-8a62-4b3a1947fa23","Type":"ContainerStarted","Data":"9b8556cdb52f6908c4d1ce583cdfbca555807483b1e8c1cb10dd9d6c724ee86b"} Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.911875 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" event={"ID":"a280b74a-cdbf-430b-9686-3055aaebf27e","Type":"ContainerStarted","Data":"3554a7c53fc95890d863673f987d4669b5b5018a8f17ed026dcde0978ac63eae"} Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.913549 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c69d55a-db50-4e0a-a86e-5b92f8088020-catalog-content\") pod \"redhat-marketplace-7whlt\" (UID: \"1c69d55a-db50-4e0a-a86e-5b92f8088020\") " pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.913683 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c69d55a-db50-4e0a-a86e-5b92f8088020-utilities\") pod \"redhat-marketplace-7whlt\" (UID: \"1c69d55a-db50-4e0a-a86e-5b92f8088020\") " pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:30 crc kubenswrapper[4872]: I0126 09:21:30.913726 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms2qj\" (UniqueName: \"kubernetes.io/projected/1c69d55a-db50-4e0a-a86e-5b92f8088020-kube-api-access-ms2qj\") pod \"redhat-marketplace-7whlt\" (UID: \"1c69d55a-db50-4e0a-a86e-5b92f8088020\") " pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:31 crc kubenswrapper[4872]: I0126 09:21:31.015379 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c69d55a-db50-4e0a-a86e-5b92f8088020-utilities\") pod \"redhat-marketplace-7whlt\" (UID: \"1c69d55a-db50-4e0a-a86e-5b92f8088020\") " pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:31 crc kubenswrapper[4872]: I0126 09:21:31.015432 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms2qj\" (UniqueName: \"kubernetes.io/projected/1c69d55a-db50-4e0a-a86e-5b92f8088020-kube-api-access-ms2qj\") pod \"redhat-marketplace-7whlt\" (UID: \"1c69d55a-db50-4e0a-a86e-5b92f8088020\") " pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:31 crc kubenswrapper[4872]: I0126 09:21:31.015484 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c69d55a-db50-4e0a-a86e-5b92f8088020-catalog-content\") pod \"redhat-marketplace-7whlt\" (UID: \"1c69d55a-db50-4e0a-a86e-5b92f8088020\") " pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:31 crc kubenswrapper[4872]: I0126 09:21:31.015979 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c69d55a-db50-4e0a-a86e-5b92f8088020-utilities\") pod \"redhat-marketplace-7whlt\" (UID: \"1c69d55a-db50-4e0a-a86e-5b92f8088020\") " pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:31 crc kubenswrapper[4872]: I0126 09:21:31.016006 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c69d55a-db50-4e0a-a86e-5b92f8088020-catalog-content\") pod \"redhat-marketplace-7whlt\" (UID: \"1c69d55a-db50-4e0a-a86e-5b92f8088020\") " pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:31 crc kubenswrapper[4872]: I0126 09:21:31.032516 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms2qj\" (UniqueName: \"kubernetes.io/projected/1c69d55a-db50-4e0a-a86e-5b92f8088020-kube-api-access-ms2qj\") pod \"redhat-marketplace-7whlt\" (UID: \"1c69d55a-db50-4e0a-a86e-5b92f8088020\") " pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:31 crc kubenswrapper[4872]: I0126 09:21:31.166403 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:31 crc kubenswrapper[4872]: I0126 09:21:31.503648 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7whlt"] Jan 26 09:21:31 crc kubenswrapper[4872]: I0126 09:21:31.931825 4872 generic.go:334] "Generic (PLEG): container finished" podID="1c69d55a-db50-4e0a-a86e-5b92f8088020" containerID="e3088861535ac5f2a9f0a42967a2c7826bf5c5e9f07ed485a2c21d305d8b560e" exitCode=0 Jan 26 09:21:31 crc kubenswrapper[4872]: I0126 09:21:31.931881 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7whlt" event={"ID":"1c69d55a-db50-4e0a-a86e-5b92f8088020","Type":"ContainerDied","Data":"e3088861535ac5f2a9f0a42967a2c7826bf5c5e9f07ed485a2c21d305d8b560e"} Jan 26 09:21:31 crc kubenswrapper[4872]: I0126 09:21:31.931939 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7whlt" event={"ID":"1c69d55a-db50-4e0a-a86e-5b92f8088020","Type":"ContainerStarted","Data":"d3f04fa747680a17d304de63789b992a2ed54db5384a9bb2eab7c643f8878e80"} Jan 26 09:21:32 crc kubenswrapper[4872]: I0126 09:21:32.713647 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-q4wxs"] Jan 26 09:21:32 crc kubenswrapper[4872]: I0126 09:21:32.753506 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cb2wp"] Jan 26 09:21:32 crc kubenswrapper[4872]: I0126 09:21:32.755302 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:21:32 crc kubenswrapper[4872]: I0126 09:21:32.765247 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cb2wp"] Jan 26 09:21:32 crc kubenswrapper[4872]: I0126 09:21:32.855832 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-dns-svc\") pod \"dnsmasq-dns-666b6646f7-cb2wp\" (UID: \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\") " pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:21:32 crc kubenswrapper[4872]: I0126 09:21:32.855942 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-config\") pod \"dnsmasq-dns-666b6646f7-cb2wp\" (UID: \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\") " pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:21:32 crc kubenswrapper[4872]: I0126 09:21:32.855963 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6f254\" (UniqueName: \"kubernetes.io/projected/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-kube-api-access-6f254\") pod \"dnsmasq-dns-666b6646f7-cb2wp\" (UID: \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\") " pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:21:32 crc kubenswrapper[4872]: I0126 09:21:32.966479 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-dns-svc\") pod \"dnsmasq-dns-666b6646f7-cb2wp\" (UID: \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\") " pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:21:32 crc kubenswrapper[4872]: I0126 09:21:32.966657 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-config\") pod \"dnsmasq-dns-666b6646f7-cb2wp\" (UID: \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\") " pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:21:32 crc kubenswrapper[4872]: I0126 09:21:32.966685 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6f254\" (UniqueName: \"kubernetes.io/projected/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-kube-api-access-6f254\") pod \"dnsmasq-dns-666b6646f7-cb2wp\" (UID: \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\") " pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:21:32 crc kubenswrapper[4872]: I0126 09:21:32.968115 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-dns-svc\") pod \"dnsmasq-dns-666b6646f7-cb2wp\" (UID: \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\") " pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:21:32 crc kubenswrapper[4872]: I0126 09:21:32.968875 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-config\") pod \"dnsmasq-dns-666b6646f7-cb2wp\" (UID: \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\") " pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.002205 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6f254\" (UniqueName: \"kubernetes.io/projected/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-kube-api-access-6f254\") pod \"dnsmasq-dns-666b6646f7-cb2wp\" (UID: \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\") " pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.057106 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mfc4j"] Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.081603 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-k8xxm"] Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.083340 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.087264 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.096575 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-k8xxm"] Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.169424 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-config\") pod \"dnsmasq-dns-57d769cc4f-k8xxm\" (UID: \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.169537 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-k8xxm\" (UID: \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.169576 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttxpx\" (UniqueName: \"kubernetes.io/projected/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-kube-api-access-ttxpx\") pod \"dnsmasq-dns-57d769cc4f-k8xxm\" (UID: \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.271573 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-k8xxm\" (UID: \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.271674 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttxpx\" (UniqueName: \"kubernetes.io/projected/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-kube-api-access-ttxpx\") pod \"dnsmasq-dns-57d769cc4f-k8xxm\" (UID: \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.271821 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-config\") pod \"dnsmasq-dns-57d769cc4f-k8xxm\" (UID: \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.273716 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-k8xxm\" (UID: \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.274144 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-config\") pod \"dnsmasq-dns-57d769cc4f-k8xxm\" (UID: \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.303409 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttxpx\" (UniqueName: \"kubernetes.io/projected/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-kube-api-access-ttxpx\") pod \"dnsmasq-dns-57d769cc4f-k8xxm\" (UID: \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\") " pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.412196 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.892967 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.895093 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.899102 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.899329 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.899501 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.899665 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5sngq" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.901375 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.907126 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.907422 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.919151 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.961870 4872 generic.go:334] "Generic (PLEG): container finished" podID="1c69d55a-db50-4e0a-a86e-5b92f8088020" containerID="5a27d6df8e7ba3107ecf3c87dc54e3932207ecd21fdd2296bc0f5c0e260dfa1c" exitCode=0 Jan 26 09:21:33 crc kubenswrapper[4872]: I0126 09:21:33.962273 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7whlt" event={"ID":"1c69d55a-db50-4e0a-a86e-5b92f8088020","Type":"ContainerDied","Data":"5a27d6df8e7ba3107ecf3c87dc54e3932207ecd21fdd2296bc0f5c0e260dfa1c"} Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:33.999891 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82c8133d-d299-4c99-b8a0-96099e834e8c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:33.999955 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82c8133d-d299-4c99-b8a0-96099e834e8c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.000252 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.000309 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.000407 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-config-data\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.000428 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.000460 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.000483 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.000502 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.000525 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.000594 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw2k5\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-kube-api-access-xw2k5\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: W0126 09:21:34.009009 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod130ac93b_f4db_4e8b_9637_f05cda3d1f0c.slice/crio-d737b945610968488bae37de443696b658df73f5111bd538347e7353e1cac604 WatchSource:0}: Error finding container d737b945610968488bae37de443696b658df73f5111bd538347e7353e1cac604: Status 404 returned error can't find the container with id d737b945610968488bae37de443696b658df73f5111bd538347e7353e1cac604 Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.013292 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-k8xxm"] Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.074984 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cb2wp"] Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.102396 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.102449 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.102506 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-config-data\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.102526 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.102548 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.102564 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.102582 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.102598 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.102642 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xw2k5\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-kube-api-access-xw2k5\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.102665 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82c8133d-d299-4c99-b8a0-96099e834e8c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.102687 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82c8133d-d299-4c99-b8a0-96099e834e8c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.104327 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.104452 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.104520 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.105403 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.105664 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.107767 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-config-data\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.110507 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82c8133d-d299-4c99-b8a0-96099e834e8c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.110527 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82c8133d-d299-4c99-b8a0-96099e834e8c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.110574 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.122010 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.123643 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw2k5\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-kube-api-access-xw2k5\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.126701 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.272994 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.281423 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.284408 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.291309 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.291567 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.291691 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.291832 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.291681 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.291995 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-grxrk" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.292127 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.292262 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.469829 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.469898 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.470027 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.470064 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.470114 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.470141 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc9lx\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-kube-api-access-xc9lx\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.470178 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.470210 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.470236 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.470322 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.470362 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.571407 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.571462 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.571488 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.571506 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.571542 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.571558 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc9lx\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-kube-api-access-xc9lx\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.571583 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.571608 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.571628 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.571664 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.571692 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.572766 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.578473 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.579001 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.579564 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.580122 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.582270 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.583453 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.585565 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.590886 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.590955 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.598644 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc9lx\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-kube-api-access-xc9lx\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.605066 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.611851 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:21:34 crc kubenswrapper[4872]: I0126 09:21:34.893203 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:34.999705 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" event={"ID":"130ac93b-f4db-4e8b-9637-f05cda3d1f0c","Type":"ContainerStarted","Data":"d737b945610968488bae37de443696b658df73f5111bd538347e7353e1cac604"} Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.007071 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" event={"ID":"35971ebc-86cc-4a97-b3ba-ebe63c6f1200","Type":"ContainerStarted","Data":"092c4fd9d43f568d59845a515ef487ab4fe967d46d850b6b56cd1d948366a5a7"} Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.023151 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"82c8133d-d299-4c99-b8a0-96099e834e8c","Type":"ContainerStarted","Data":"591c59aa46056780153b6968955e58429d6684e7d18c90693cde199497de5c5f"} Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.386112 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.426892 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Jan 26 09:21:35 crc kubenswrapper[4872]: W0126 09:21:35.448463 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0792ad43_38af_4c94_bd3b_fc7d3a7af5da.slice/crio-5f771a049c7d1063e46471249c3a8b5782df08fa31a10153c65849443943a9a4 WatchSource:0}: Error finding container 5f771a049c7d1063e46471249c3a8b5782df08fa31a10153c65849443943a9a4: Status 404 returned error can't find the container with id 5f771a049c7d1063e46471249c3a8b5782df08fa31a10153c65849443943a9a4 Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.457308 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.457480 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.460133 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.460681 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.461699 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-9ffhr" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.461954 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.468562 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.590377 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.590436 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9082961a-1e93-49b5-8420-7516b675ea6f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.590461 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9082961a-1e93-49b5-8420-7516b675ea6f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.590501 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9082961a-1e93-49b5-8420-7516b675ea6f-config-data-default\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.590572 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9082961a-1e93-49b5-8420-7516b675ea6f-kolla-config\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.590590 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9082961a-1e93-49b5-8420-7516b675ea6f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.590697 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvz9d\" (UniqueName: \"kubernetes.io/projected/9082961a-1e93-49b5-8420-7516b675ea6f-kube-api-access-kvz9d\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.590765 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9082961a-1e93-49b5-8420-7516b675ea6f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.695779 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.695909 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9082961a-1e93-49b5-8420-7516b675ea6f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.695961 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9082961a-1e93-49b5-8420-7516b675ea6f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.696047 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9082961a-1e93-49b5-8420-7516b675ea6f-config-data-default\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.696132 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9082961a-1e93-49b5-8420-7516b675ea6f-kolla-config\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.696158 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9082961a-1e93-49b5-8420-7516b675ea6f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.696215 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvz9d\" (UniqueName: \"kubernetes.io/projected/9082961a-1e93-49b5-8420-7516b675ea6f-kube-api-access-kvz9d\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.696250 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9082961a-1e93-49b5-8420-7516b675ea6f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.698041 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9082961a-1e93-49b5-8420-7516b675ea6f-config-data-default\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.698154 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9082961a-1e93-49b5-8420-7516b675ea6f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.698555 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.702881 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9082961a-1e93-49b5-8420-7516b675ea6f-kolla-config\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.703708 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9082961a-1e93-49b5-8420-7516b675ea6f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.721570 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9082961a-1e93-49b5-8420-7516b675ea6f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.722356 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9082961a-1e93-49b5-8420-7516b675ea6f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.746619 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvz9d\" (UniqueName: \"kubernetes.io/projected/9082961a-1e93-49b5-8420-7516b675ea6f-kube-api-access-kvz9d\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:35 crc kubenswrapper[4872]: I0126 09:21:35.834682 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"9082961a-1e93-49b5-8420-7516b675ea6f\") " pod="openstack/openstack-galera-0" Jan 26 09:21:36 crc kubenswrapper[4872]: I0126 09:21:36.067734 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0792ad43-38af-4c94-bd3b-fc7d3a7af5da","Type":"ContainerStarted","Data":"5f771a049c7d1063e46471249c3a8b5782df08fa31a10153c65849443943a9a4"} Jan 26 09:21:36 crc kubenswrapper[4872]: I0126 09:21:36.073421 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7whlt" event={"ID":"1c69d55a-db50-4e0a-a86e-5b92f8088020","Type":"ContainerStarted","Data":"22da54cf95d21aac3921ecfc89b0b8fee6d16dfe71212c4fe0d2bc9c81833cd9"} Jan 26 09:21:36 crc kubenswrapper[4872]: I0126 09:21:36.099364 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Jan 26 09:21:36 crc kubenswrapper[4872]: I0126 09:21:36.132583 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7whlt" podStartSLOduration=3.423906252 podStartE2EDuration="6.132564315s" podCreationTimestamp="2026-01-26 09:21:30 +0000 UTC" firstStartedPulling="2026-01-26 09:21:31.935048858 +0000 UTC m=+825.243888659" lastFinishedPulling="2026-01-26 09:21:34.643706921 +0000 UTC m=+827.952546722" observedRunningTime="2026-01-26 09:21:36.129703393 +0000 UTC m=+829.438543214" watchObservedRunningTime="2026-01-26 09:21:36.132564315 +0000 UTC m=+829.441404116" Jan 26 09:21:36 crc kubenswrapper[4872]: I0126 09:21:36.812353 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Jan 26 09:21:36 crc kubenswrapper[4872]: I0126 09:21:36.947989 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 26 09:21:36 crc kubenswrapper[4872]: I0126 09:21:36.957784 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:36 crc kubenswrapper[4872]: I0126 09:21:36.962072 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Jan 26 09:21:36 crc kubenswrapper[4872]: I0126 09:21:36.962455 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Jan 26 09:21:36 crc kubenswrapper[4872]: I0126 09:21:36.962704 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-tdqff" Jan 26 09:21:36 crc kubenswrapper[4872]: I0126 09:21:36.962883 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Jan 26 09:21:36 crc kubenswrapper[4872]: I0126 09:21:36.978911 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.105582 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.108755 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.117348 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.117749 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.128224 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-tkjds" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.144142 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.172455 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ab5ce366-26e9-424a-9a73-9173636c2902-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.172945 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab5ce366-26e9-424a-9a73-9173636c2902-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.173097 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ab5ce366-26e9-424a-9a73-9173636c2902-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.173165 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5ce366-26e9-424a-9a73-9173636c2902-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.173346 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rszfn\" (UniqueName: \"kubernetes.io/projected/ab5ce366-26e9-424a-9a73-9173636c2902-kube-api-access-rszfn\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.173404 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.173471 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ab5ce366-26e9-424a-9a73-9173636c2902-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.173563 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab5ce366-26e9-424a-9a73-9173636c2902-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.258870 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9082961a-1e93-49b5-8420-7516b675ea6f","Type":"ContainerStarted","Data":"e31f2d578c36fdca8150a633fe582b00648155472604cbae7e1c38e56909b9d4"} Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285264 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab5ce366-26e9-424a-9a73-9173636c2902-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285312 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98a9098c-4abc-4a20-be26-d62bd842bd04-config-data\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285347 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/98a9098c-4abc-4a20-be26-d62bd842bd04-kolla-config\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285381 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnj96\" (UniqueName: \"kubernetes.io/projected/98a9098c-4abc-4a20-be26-d62bd842bd04-kube-api-access-pnj96\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285409 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ab5ce366-26e9-424a-9a73-9173636c2902-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285435 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a9098c-4abc-4a20-be26-d62bd842bd04-combined-ca-bundle\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285459 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab5ce366-26e9-424a-9a73-9173636c2902-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285481 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ab5ce366-26e9-424a-9a73-9173636c2902-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285500 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5ce366-26e9-424a-9a73-9173636c2902-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285525 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/98a9098c-4abc-4a20-be26-d62bd842bd04-memcached-tls-certs\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285555 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rszfn\" (UniqueName: \"kubernetes.io/projected/ab5ce366-26e9-424a-9a73-9173636c2902-kube-api-access-rszfn\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285575 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.285595 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ab5ce366-26e9-424a-9a73-9173636c2902-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.291003 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ab5ce366-26e9-424a-9a73-9173636c2902-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.292407 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ab5ce366-26e9-424a-9a73-9173636c2902-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.297265 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ab5ce366-26e9-424a-9a73-9173636c2902-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.299166 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ab5ce366-26e9-424a-9a73-9173636c2902-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.301438 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.323411 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab5ce366-26e9-424a-9a73-9173636c2902-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.323497 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab5ce366-26e9-424a-9a73-9173636c2902-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.333778 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rszfn\" (UniqueName: \"kubernetes.io/projected/ab5ce366-26e9-424a-9a73-9173636c2902-kube-api-access-rszfn\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.361137 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"ab5ce366-26e9-424a-9a73-9173636c2902\") " pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.409413 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/98a9098c-4abc-4a20-be26-d62bd842bd04-memcached-tls-certs\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.409499 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98a9098c-4abc-4a20-be26-d62bd842bd04-config-data\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.409540 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/98a9098c-4abc-4a20-be26-d62bd842bd04-kolla-config\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.409565 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnj96\" (UniqueName: \"kubernetes.io/projected/98a9098c-4abc-4a20-be26-d62bd842bd04-kube-api-access-pnj96\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.409602 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a9098c-4abc-4a20-be26-d62bd842bd04-combined-ca-bundle\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.410833 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98a9098c-4abc-4a20-be26-d62bd842bd04-config-data\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.414477 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/98a9098c-4abc-4a20-be26-d62bd842bd04-memcached-tls-certs\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.415040 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/98a9098c-4abc-4a20-be26-d62bd842bd04-kolla-config\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.419328 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98a9098c-4abc-4a20-be26-d62bd842bd04-combined-ca-bundle\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.444546 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnj96\" (UniqueName: \"kubernetes.io/projected/98a9098c-4abc-4a20-be26-d62bd842bd04-kube-api-access-pnj96\") pod \"memcached-0\" (UID: \"98a9098c-4abc-4a20-be26-d62bd842bd04\") " pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.448250 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Jan 26 09:21:37 crc kubenswrapper[4872]: I0126 09:21:37.585229 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Jan 26 09:21:38 crc kubenswrapper[4872]: I0126 09:21:38.061664 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Jan 26 09:21:38 crc kubenswrapper[4872]: I0126 09:21:38.210190 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"98a9098c-4abc-4a20-be26-d62bd842bd04","Type":"ContainerStarted","Data":"9a553cd0937cdde587d926a2588ea3ec40841fd0f8c3d065b186fc93c767361b"} Jan 26 09:21:38 crc kubenswrapper[4872]: I0126 09:21:38.451005 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Jan 26 09:21:39 crc kubenswrapper[4872]: I0126 09:21:39.276769 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ab5ce366-26e9-424a-9a73-9173636c2902","Type":"ContainerStarted","Data":"8e97827c485f74f77c3922609df01df4661463162d1e36de2fbb7565cc8fbb2f"} Jan 26 09:21:39 crc kubenswrapper[4872]: I0126 09:21:39.497538 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 09:21:39 crc kubenswrapper[4872]: I0126 09:21:39.500657 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 26 09:21:39 crc kubenswrapper[4872]: I0126 09:21:39.508187 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-c7cm8" Jan 26 09:21:39 crc kubenswrapper[4872]: I0126 09:21:39.512099 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 09:21:39 crc kubenswrapper[4872]: I0126 09:21:39.590048 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p98pt\" (UniqueName: \"kubernetes.io/projected/3d16170b-8fc9-4c93-85fe-9ccef4da5bd0-kube-api-access-p98pt\") pod \"kube-state-metrics-0\" (UID: \"3d16170b-8fc9-4c93-85fe-9ccef4da5bd0\") " pod="openstack/kube-state-metrics-0" Jan 26 09:21:39 crc kubenswrapper[4872]: I0126 09:21:39.692084 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p98pt\" (UniqueName: \"kubernetes.io/projected/3d16170b-8fc9-4c93-85fe-9ccef4da5bd0-kube-api-access-p98pt\") pod \"kube-state-metrics-0\" (UID: \"3d16170b-8fc9-4c93-85fe-9ccef4da5bd0\") " pod="openstack/kube-state-metrics-0" Jan 26 09:21:39 crc kubenswrapper[4872]: I0126 09:21:39.725769 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p98pt\" (UniqueName: \"kubernetes.io/projected/3d16170b-8fc9-4c93-85fe-9ccef4da5bd0-kube-api-access-p98pt\") pod \"kube-state-metrics-0\" (UID: \"3d16170b-8fc9-4c93-85fe-9ccef4da5bd0\") " pod="openstack/kube-state-metrics-0" Jan 26 09:21:39 crc kubenswrapper[4872]: I0126 09:21:39.848780 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 26 09:21:40 crc kubenswrapper[4872]: I0126 09:21:40.592369 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.167039 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.167578 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.245378 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.364649 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3d16170b-8fc9-4c93-85fe-9ccef4da5bd0","Type":"ContainerStarted","Data":"94e7d545c7f6e0ad17da0f7fe0de17d2c79b872267a79d0348cf739b5ce3ee7c"} Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.478827 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.528850 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bkdtq"] Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.535286 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.566961 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bkdtq"] Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.658282 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e318d821-9c39-43ac-92e7-8834d9275c06-utilities\") pod \"certified-operators-bkdtq\" (UID: \"e318d821-9c39-43ac-92e7-8834d9275c06\") " pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.658346 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e318d821-9c39-43ac-92e7-8834d9275c06-catalog-content\") pod \"certified-operators-bkdtq\" (UID: \"e318d821-9c39-43ac-92e7-8834d9275c06\") " pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.658403 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh6bp\" (UniqueName: \"kubernetes.io/projected/e318d821-9c39-43ac-92e7-8834d9275c06-kube-api-access-hh6bp\") pod \"certified-operators-bkdtq\" (UID: \"e318d821-9c39-43ac-92e7-8834d9275c06\") " pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.760621 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e318d821-9c39-43ac-92e7-8834d9275c06-utilities\") pod \"certified-operators-bkdtq\" (UID: \"e318d821-9c39-43ac-92e7-8834d9275c06\") " pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.760699 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e318d821-9c39-43ac-92e7-8834d9275c06-catalog-content\") pod \"certified-operators-bkdtq\" (UID: \"e318d821-9c39-43ac-92e7-8834d9275c06\") " pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.760741 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh6bp\" (UniqueName: \"kubernetes.io/projected/e318d821-9c39-43ac-92e7-8834d9275c06-kube-api-access-hh6bp\") pod \"certified-operators-bkdtq\" (UID: \"e318d821-9c39-43ac-92e7-8834d9275c06\") " pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.761770 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e318d821-9c39-43ac-92e7-8834d9275c06-utilities\") pod \"certified-operators-bkdtq\" (UID: \"e318d821-9c39-43ac-92e7-8834d9275c06\") " pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.767731 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e318d821-9c39-43ac-92e7-8834d9275c06-catalog-content\") pod \"certified-operators-bkdtq\" (UID: \"e318d821-9c39-43ac-92e7-8834d9275c06\") " pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.830850 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh6bp\" (UniqueName: \"kubernetes.io/projected/e318d821-9c39-43ac-92e7-8834d9275c06-kube-api-access-hh6bp\") pod \"certified-operators-bkdtq\" (UID: \"e318d821-9c39-43ac-92e7-8834d9275c06\") " pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:21:41 crc kubenswrapper[4872]: I0126 09:21:41.934437 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.776109 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gmp2c"] Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.778107 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.782452 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.782692 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.786509 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-bzw55" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.813650 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gmp2c"] Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.873921 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-mgn5r"] Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.878931 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.882959 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mgn5r"] Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.925818 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4ece36ce-cbb6-4012-8b5c-87834502f26b-etc-ovs\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.926414 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4ece36ce-cbb6-4012-8b5c-87834502f26b-var-log\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.926472 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cb72fdef-372c-41a1-bf57-6c71162ab194-var-run\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.926497 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cb72fdef-372c-41a1-bf57-6c71162ab194-var-run-ovn\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.926534 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ece36ce-cbb6-4012-8b5c-87834502f26b-scripts\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.926729 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cb72fdef-372c-41a1-bf57-6c71162ab194-var-log-ovn\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.926779 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg2z7\" (UniqueName: \"kubernetes.io/projected/cb72fdef-372c-41a1-bf57-6c71162ab194-kube-api-access-pg2z7\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.926926 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb72fdef-372c-41a1-bf57-6c71162ab194-combined-ca-bundle\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.927235 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgzz8\" (UniqueName: \"kubernetes.io/projected/4ece36ce-cbb6-4012-8b5c-87834502f26b-kube-api-access-lgzz8\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.927280 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4ece36ce-cbb6-4012-8b5c-87834502f26b-var-lib\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.927331 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4ece36ce-cbb6-4012-8b5c-87834502f26b-var-run\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.927364 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb72fdef-372c-41a1-bf57-6c71162ab194-ovn-controller-tls-certs\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.927542 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb72fdef-372c-41a1-bf57-6c71162ab194-scripts\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:42 crc kubenswrapper[4872]: I0126 09:21:42.961023 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bkdtq"] Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.029608 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cb72fdef-372c-41a1-bf57-6c71162ab194-var-run\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.029676 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cb72fdef-372c-41a1-bf57-6c71162ab194-var-run-ovn\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.029702 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4ece36ce-cbb6-4012-8b5c-87834502f26b-var-log\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.029730 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ece36ce-cbb6-4012-8b5c-87834502f26b-scripts\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.031081 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cb72fdef-372c-41a1-bf57-6c71162ab194-var-log-ovn\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.031158 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg2z7\" (UniqueName: \"kubernetes.io/projected/cb72fdef-372c-41a1-bf57-6c71162ab194-kube-api-access-pg2z7\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.031182 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb72fdef-372c-41a1-bf57-6c71162ab194-combined-ca-bundle\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.031248 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgzz8\" (UniqueName: \"kubernetes.io/projected/4ece36ce-cbb6-4012-8b5c-87834502f26b-kube-api-access-lgzz8\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.031272 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4ece36ce-cbb6-4012-8b5c-87834502f26b-var-lib\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.031326 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb72fdef-372c-41a1-bf57-6c71162ab194-ovn-controller-tls-certs\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.031344 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4ece36ce-cbb6-4012-8b5c-87834502f26b-var-run\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.031468 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb72fdef-372c-41a1-bf57-6c71162ab194-scripts\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.031505 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4ece36ce-cbb6-4012-8b5c-87834502f26b-etc-ovs\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.031856 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ece36ce-cbb6-4012-8b5c-87834502f26b-scripts\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.032247 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cb72fdef-372c-41a1-bf57-6c71162ab194-var-run\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.032715 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cb72fdef-372c-41a1-bf57-6c71162ab194-var-log-ovn\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.036066 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4ece36ce-cbb6-4012-8b5c-87834502f26b-var-log\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.036357 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4ece36ce-cbb6-4012-8b5c-87834502f26b-etc-ovs\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.036562 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4ece36ce-cbb6-4012-8b5c-87834502f26b-var-lib\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.037033 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4ece36ce-cbb6-4012-8b5c-87834502f26b-var-run\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.038385 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cb72fdef-372c-41a1-bf57-6c71162ab194-scripts\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.042602 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cb72fdef-372c-41a1-bf57-6c71162ab194-ovn-controller-tls-certs\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.042853 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cb72fdef-372c-41a1-bf57-6c71162ab194-combined-ca-bundle\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.052080 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cb72fdef-372c-41a1-bf57-6c71162ab194-var-run-ovn\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.070695 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg2z7\" (UniqueName: \"kubernetes.io/projected/cb72fdef-372c-41a1-bf57-6c71162ab194-kube-api-access-pg2z7\") pod \"ovn-controller-gmp2c\" (UID: \"cb72fdef-372c-41a1-bf57-6c71162ab194\") " pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.095308 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgzz8\" (UniqueName: \"kubernetes.io/projected/4ece36ce-cbb6-4012-8b5c-87834502f26b-kube-api-access-lgzz8\") pod \"ovn-controller-ovs-mgn5r\" (UID: \"4ece36ce-cbb6-4012-8b5c-87834502f26b\") " pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.106744 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gmp2c" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.218609 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.397414 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkdtq" event={"ID":"e318d821-9c39-43ac-92e7-8834d9275c06","Type":"ContainerStarted","Data":"e9d7dc247b33851f470f99549e1670016adcb90c412e43b62326b06e8cb7047f"} Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.671722 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.696856 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.704676 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.705037 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-2bxjf" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.705192 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.705505 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.705625 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.706289 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.743346 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7whlt"] Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.743640 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7whlt" podUID="1c69d55a-db50-4e0a-a86e-5b92f8088020" containerName="registry-server" containerID="cri-o://22da54cf95d21aac3921ecfc89b0b8fee6d16dfe71212c4fe0d2bc9c81833cd9" gracePeriod=2 Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.859655 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c56wm\" (UniqueName: \"kubernetes.io/projected/bf0544f2-34dc-45fc-986e-cb2554c04566-kube-api-access-c56wm\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.859736 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf0544f2-34dc-45fc-986e-cb2554c04566-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.859775 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bf0544f2-34dc-45fc-986e-cb2554c04566-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.859846 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf0544f2-34dc-45fc-986e-cb2554c04566-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.860079 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.860280 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf0544f2-34dc-45fc-986e-cb2554c04566-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.860427 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0544f2-34dc-45fc-986e-cb2554c04566-config\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.861409 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0544f2-34dc-45fc-986e-cb2554c04566-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.963096 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf0544f2-34dc-45fc-986e-cb2554c04566-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.963167 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.963200 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf0544f2-34dc-45fc-986e-cb2554c04566-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.963245 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0544f2-34dc-45fc-986e-cb2554c04566-config\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.963281 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0544f2-34dc-45fc-986e-cb2554c04566-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.963337 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c56wm\" (UniqueName: \"kubernetes.io/projected/bf0544f2-34dc-45fc-986e-cb2554c04566-kube-api-access-c56wm\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.963374 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf0544f2-34dc-45fc-986e-cb2554c04566-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.963408 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bf0544f2-34dc-45fc-986e-cb2554c04566-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.963990 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bf0544f2-34dc-45fc-986e-cb2554c04566-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.965712 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf0544f2-34dc-45fc-986e-cb2554c04566-config\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.966023 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.967521 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf0544f2-34dc-45fc-986e-cb2554c04566-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.974728 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf0544f2-34dc-45fc-986e-cb2554c04566-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.974966 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf0544f2-34dc-45fc-986e-cb2554c04566-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:43 crc kubenswrapper[4872]: I0126 09:21:43.998682 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf0544f2-34dc-45fc-986e-cb2554c04566-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:44 crc kubenswrapper[4872]: I0126 09:21:44.004605 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c56wm\" (UniqueName: \"kubernetes.io/projected/bf0544f2-34dc-45fc-986e-cb2554c04566-kube-api-access-c56wm\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:44 crc kubenswrapper[4872]: I0126 09:21:44.008364 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ovsdbserver-nb-0\" (UID: \"bf0544f2-34dc-45fc-986e-cb2554c04566\") " pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:44 crc kubenswrapper[4872]: I0126 09:21:44.054830 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Jan 26 09:21:44 crc kubenswrapper[4872]: I0126 09:21:44.414492 4872 generic.go:334] "Generic (PLEG): container finished" podID="1c69d55a-db50-4e0a-a86e-5b92f8088020" containerID="22da54cf95d21aac3921ecfc89b0b8fee6d16dfe71212c4fe0d2bc9c81833cd9" exitCode=0 Jan 26 09:21:44 crc kubenswrapper[4872]: I0126 09:21:44.414547 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7whlt" event={"ID":"1c69d55a-db50-4e0a-a86e-5b92f8088020","Type":"ContainerDied","Data":"22da54cf95d21aac3921ecfc89b0b8fee6d16dfe71212c4fe0d2bc9c81833cd9"} Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.760844 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.763048 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.767636 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-mdx2k" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.768668 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.770017 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.775366 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.786730 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.869768 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/204a2939-a385-4f7d-a08b-a008e6a14df0-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.870264 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.870288 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/204a2939-a385-4f7d-a08b-a008e6a14df0-config\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.870320 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/204a2939-a385-4f7d-a08b-a008e6a14df0-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.870340 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/204a2939-a385-4f7d-a08b-a008e6a14df0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.870371 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/204a2939-a385-4f7d-a08b-a008e6a14df0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.870411 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ssz5\" (UniqueName: \"kubernetes.io/projected/204a2939-a385-4f7d-a08b-a008e6a14df0-kube-api-access-4ssz5\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.870447 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/204a2939-a385-4f7d-a08b-a008e6a14df0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.971691 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/204a2939-a385-4f7d-a08b-a008e6a14df0-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.971746 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/204a2939-a385-4f7d-a08b-a008e6a14df0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.971780 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/204a2939-a385-4f7d-a08b-a008e6a14df0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.971833 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ssz5\" (UniqueName: \"kubernetes.io/projected/204a2939-a385-4f7d-a08b-a008e6a14df0-kube-api-access-4ssz5\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.971871 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/204a2939-a385-4f7d-a08b-a008e6a14df0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.971927 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/204a2939-a385-4f7d-a08b-a008e6a14df0-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.971948 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.971968 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/204a2939-a385-4f7d-a08b-a008e6a14df0-config\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.972320 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/204a2939-a385-4f7d-a08b-a008e6a14df0-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.972471 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.972843 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/204a2939-a385-4f7d-a08b-a008e6a14df0-config\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.973930 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/204a2939-a385-4f7d-a08b-a008e6a14df0-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.984119 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/204a2939-a385-4f7d-a08b-a008e6a14df0-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.986629 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/204a2939-a385-4f7d-a08b-a008e6a14df0-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.993631 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ssz5\" (UniqueName: \"kubernetes.io/projected/204a2939-a385-4f7d-a08b-a008e6a14df0-kube-api-access-4ssz5\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:46 crc kubenswrapper[4872]: I0126 09:21:46.996756 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/204a2939-a385-4f7d-a08b-a008e6a14df0-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:47 crc kubenswrapper[4872]: I0126 09:21:47.007257 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-sb-0\" (UID: \"204a2939-a385-4f7d-a08b-a008e6a14df0\") " pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:47 crc kubenswrapper[4872]: I0126 09:21:47.089310 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Jan 26 09:21:51 crc kubenswrapper[4872]: E0126 09:21:51.167573 4872 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 22da54cf95d21aac3921ecfc89b0b8fee6d16dfe71212c4fe0d2bc9c81833cd9 is running failed: container process not found" containerID="22da54cf95d21aac3921ecfc89b0b8fee6d16dfe71212c4fe0d2bc9c81833cd9" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 09:21:51 crc kubenswrapper[4872]: E0126 09:21:51.170134 4872 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 22da54cf95d21aac3921ecfc89b0b8fee6d16dfe71212c4fe0d2bc9c81833cd9 is running failed: container process not found" containerID="22da54cf95d21aac3921ecfc89b0b8fee6d16dfe71212c4fe0d2bc9c81833cd9" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 09:21:51 crc kubenswrapper[4872]: E0126 09:21:51.170653 4872 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 22da54cf95d21aac3921ecfc89b0b8fee6d16dfe71212c4fe0d2bc9c81833cd9 is running failed: container process not found" containerID="22da54cf95d21aac3921ecfc89b0b8fee6d16dfe71212c4fe0d2bc9c81833cd9" cmd=["grpc_health_probe","-addr=:50051"] Jan 26 09:21:51 crc kubenswrapper[4872]: E0126 09:21:51.170696 4872 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 22da54cf95d21aac3921ecfc89b0b8fee6d16dfe71212c4fe0d2bc9c81833cd9 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-7whlt" podUID="1c69d55a-db50-4e0a-a86e-5b92f8088020" containerName="registry-server" Jan 26 09:21:53 crc kubenswrapper[4872]: I0126 09:21:53.903328 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dmpdv"] Jan 26 09:21:53 crc kubenswrapper[4872]: I0126 09:21:53.906069 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:21:53 crc kubenswrapper[4872]: I0126 09:21:53.913395 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dmpdv"] Jan 26 09:21:53 crc kubenswrapper[4872]: I0126 09:21:53.932946 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/179dc633-3c59-4816-9ddf-db909211be5c-catalog-content\") pod \"redhat-operators-dmpdv\" (UID: \"179dc633-3c59-4816-9ddf-db909211be5c\") " pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:21:53 crc kubenswrapper[4872]: I0126 09:21:53.933007 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/179dc633-3c59-4816-9ddf-db909211be5c-utilities\") pod \"redhat-operators-dmpdv\" (UID: \"179dc633-3c59-4816-9ddf-db909211be5c\") " pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:21:53 crc kubenswrapper[4872]: I0126 09:21:53.933049 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7wfd\" (UniqueName: \"kubernetes.io/projected/179dc633-3c59-4816-9ddf-db909211be5c-kube-api-access-k7wfd\") pod \"redhat-operators-dmpdv\" (UID: \"179dc633-3c59-4816-9ddf-db909211be5c\") " pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:21:54 crc kubenswrapper[4872]: I0126 09:21:54.034485 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/179dc633-3c59-4816-9ddf-db909211be5c-utilities\") pod \"redhat-operators-dmpdv\" (UID: \"179dc633-3c59-4816-9ddf-db909211be5c\") " pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:21:54 crc kubenswrapper[4872]: I0126 09:21:54.034551 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7wfd\" (UniqueName: \"kubernetes.io/projected/179dc633-3c59-4816-9ddf-db909211be5c-kube-api-access-k7wfd\") pod \"redhat-operators-dmpdv\" (UID: \"179dc633-3c59-4816-9ddf-db909211be5c\") " pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:21:54 crc kubenswrapper[4872]: I0126 09:21:54.034666 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/179dc633-3c59-4816-9ddf-db909211be5c-catalog-content\") pod \"redhat-operators-dmpdv\" (UID: \"179dc633-3c59-4816-9ddf-db909211be5c\") " pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:21:54 crc kubenswrapper[4872]: I0126 09:21:54.035109 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/179dc633-3c59-4816-9ddf-db909211be5c-catalog-content\") pod \"redhat-operators-dmpdv\" (UID: \"179dc633-3c59-4816-9ddf-db909211be5c\") " pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:21:54 crc kubenswrapper[4872]: I0126 09:21:54.035421 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/179dc633-3c59-4816-9ddf-db909211be5c-utilities\") pod \"redhat-operators-dmpdv\" (UID: \"179dc633-3c59-4816-9ddf-db909211be5c\") " pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:21:54 crc kubenswrapper[4872]: I0126 09:21:54.061813 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7wfd\" (UniqueName: \"kubernetes.io/projected/179dc633-3c59-4816-9ddf-db909211be5c-kube-api-access-k7wfd\") pod \"redhat-operators-dmpdv\" (UID: \"179dc633-3c59-4816-9ddf-db909211be5c\") " pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:21:54 crc kubenswrapper[4872]: I0126 09:21:54.240678 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:22:00 crc kubenswrapper[4872]: E0126 09:22:00.050251 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Jan 26 09:22:00 crc kubenswrapper[4872]: E0126 09:22:00.051478 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n56ch87h547h668h54h57bh68fh648hc6h5d6h5f8hb4h678h579h66fh546h64fh54fh599h5c6h578hfbh695hd9h687h695h65ch69h66h56bh5fch5c8q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pnj96,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(98a9098c-4abc-4a20-be26-d62bd842bd04): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:22:00 crc kubenswrapper[4872]: E0126 09:22:00.052855 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="98a9098c-4abc-4a20-be26-d62bd842bd04" Jan 26 09:22:00 crc kubenswrapper[4872]: E0126 09:22:00.606552 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="98a9098c-4abc-4a20-be26-d62bd842bd04" Jan 26 09:22:00 crc kubenswrapper[4872]: E0126 09:22:00.994309 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Jan 26 09:22:00 crc kubenswrapper[4872]: E0126 09:22:00.994571 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xc9lx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(0792ad43-38af-4c94-bd3b-fc7d3a7af5da): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:22:00 crc kubenswrapper[4872]: E0126 09:22:00.995850 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="0792ad43-38af-4c94-bd3b-fc7d3a7af5da" Jan 26 09:22:01 crc kubenswrapper[4872]: E0126 09:22:01.022292 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Jan 26 09:22:01 crc kubenswrapper[4872]: E0126 09:22:01.022764 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xw2k5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(82c8133d-d299-4c99-b8a0-96099e834e8c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:22:01 crc kubenswrapper[4872]: E0126 09:22:01.024022 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="82c8133d-d299-4c99-b8a0-96099e834e8c" Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.057702 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.068173 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c69d55a-db50-4e0a-a86e-5b92f8088020-utilities\") pod \"1c69d55a-db50-4e0a-a86e-5b92f8088020\" (UID: \"1c69d55a-db50-4e0a-a86e-5b92f8088020\") " Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.068307 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c69d55a-db50-4e0a-a86e-5b92f8088020-catalog-content\") pod \"1c69d55a-db50-4e0a-a86e-5b92f8088020\" (UID: \"1c69d55a-db50-4e0a-a86e-5b92f8088020\") " Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.068331 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms2qj\" (UniqueName: \"kubernetes.io/projected/1c69d55a-db50-4e0a-a86e-5b92f8088020-kube-api-access-ms2qj\") pod \"1c69d55a-db50-4e0a-a86e-5b92f8088020\" (UID: \"1c69d55a-db50-4e0a-a86e-5b92f8088020\") " Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.069233 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c69d55a-db50-4e0a-a86e-5b92f8088020-utilities" (OuterVolumeSpecName: "utilities") pod "1c69d55a-db50-4e0a-a86e-5b92f8088020" (UID: "1c69d55a-db50-4e0a-a86e-5b92f8088020"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.083709 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c69d55a-db50-4e0a-a86e-5b92f8088020-kube-api-access-ms2qj" (OuterVolumeSpecName: "kube-api-access-ms2qj") pod "1c69d55a-db50-4e0a-a86e-5b92f8088020" (UID: "1c69d55a-db50-4e0a-a86e-5b92f8088020"). InnerVolumeSpecName "kube-api-access-ms2qj". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.096836 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c69d55a-db50-4e0a-a86e-5b92f8088020-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c69d55a-db50-4e0a-a86e-5b92f8088020" (UID: "1c69d55a-db50-4e0a-a86e-5b92f8088020"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.169779 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c69d55a-db50-4e0a-a86e-5b92f8088020-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.169853 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c69d55a-db50-4e0a-a86e-5b92f8088020-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.169871 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms2qj\" (UniqueName: \"kubernetes.io/projected/1c69d55a-db50-4e0a-a86e-5b92f8088020-kube-api-access-ms2qj\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.618219 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7whlt" Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.618903 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7whlt" event={"ID":"1c69d55a-db50-4e0a-a86e-5b92f8088020","Type":"ContainerDied","Data":"d3f04fa747680a17d304de63789b992a2ed54db5384a9bb2eab7c643f8878e80"} Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.618949 4872 scope.go:117] "RemoveContainer" containerID="22da54cf95d21aac3921ecfc89b0b8fee6d16dfe71212c4fe0d2bc9c81833cd9" Jan 26 09:22:01 crc kubenswrapper[4872]: E0126 09:22:01.620438 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="82c8133d-d299-4c99-b8a0-96099e834e8c" Jan 26 09:22:01 crc kubenswrapper[4872]: E0126 09:22:01.620677 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="0792ad43-38af-4c94-bd3b-fc7d3a7af5da" Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.707850 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7whlt"] Jan 26 09:22:01 crc kubenswrapper[4872]: I0126 09:22:01.716909 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7whlt"] Jan 26 09:22:02 crc kubenswrapper[4872]: E0126 09:22:02.650104 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Jan 26 09:22:02 crc kubenswrapper[4872]: E0126 09:22:02.650379 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kvz9d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(9082961a-1e93-49b5-8420-7516b675ea6f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:22:02 crc kubenswrapper[4872]: E0126 09:22:02.651866 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="9082961a-1e93-49b5-8420-7516b675ea6f" Jan 26 09:22:02 crc kubenswrapper[4872]: E0126 09:22:02.682535 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Jan 26 09:22:02 crc kubenswrapper[4872]: E0126 09:22:02.682734 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rszfn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(ab5ce366-26e9-424a-9a73-9173636c2902): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:22:02 crc kubenswrapper[4872]: E0126 09:22:02.684086 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="ab5ce366-26e9-424a-9a73-9173636c2902" Jan 26 09:22:03 crc kubenswrapper[4872]: I0126 09:22:03.196285 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c69d55a-db50-4e0a-a86e-5b92f8088020" path="/var/lib/kubelet/pods/1c69d55a-db50-4e0a-a86e-5b92f8088020/volumes" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.371174 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.371464 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ttxpx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-k8xxm_openstack(130ac93b-f4db-4e8b-9637-f05cda3d1f0c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.372824 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" podUID="130ac93b-f4db-4e8b-9637-f05cda3d1f0c" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.378089 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.379007 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zstx7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-mfc4j_openstack(a280b74a-cdbf-430b-9686-3055aaebf27e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.381221 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" podUID="a280b74a-cdbf-430b-9686-3055aaebf27e" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.407844 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.408139 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6f254,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-cb2wp_openstack(35971ebc-86cc-4a97-b3ba-ebe63c6f1200): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.409442 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" podUID="35971ebc-86cc-4a97-b3ba-ebe63c6f1200" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.419548 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.419771 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-km4x2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-q4wxs_openstack(26bf0d4f-2ed3-4428-8a62-4b3a1947fa23): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.421436 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" podUID="26bf0d4f-2ed3-4428-8a62-4b3a1947fa23" Jan 26 09:22:03 crc kubenswrapper[4872]: I0126 09:22:03.467709 4872 scope.go:117] "RemoveContainer" containerID="5a27d6df8e7ba3107ecf3c87dc54e3932207ecd21fdd2296bc0f5c0e260dfa1c" Jan 26 09:22:03 crc kubenswrapper[4872]: I0126 09:22:03.511235 4872 scope.go:117] "RemoveContainer" containerID="e3088861535ac5f2a9f0a42967a2c7826bf5c5e9f07ed485a2c21d305d8b560e" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.635863 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="ab5ce366-26e9-424a-9a73-9173636c2902" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.635887 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" podUID="130ac93b-f4db-4e8b-9637-f05cda3d1f0c" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.636218 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-galera-0" podUID="9082961a-1e93-49b5-8420-7516b675ea6f" Jan 26 09:22:03 crc kubenswrapper[4872]: E0126 09:22:03.648609 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" podUID="35971ebc-86cc-4a97-b3ba-ebe63c6f1200" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.282722 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.291591 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.338427 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23-config\") pod \"26bf0d4f-2ed3-4428-8a62-4b3a1947fa23\" (UID: \"26bf0d4f-2ed3-4428-8a62-4b3a1947fa23\") " Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.338626 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km4x2\" (UniqueName: \"kubernetes.io/projected/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23-kube-api-access-km4x2\") pod \"26bf0d4f-2ed3-4428-8a62-4b3a1947fa23\" (UID: \"26bf0d4f-2ed3-4428-8a62-4b3a1947fa23\") " Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.338706 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zstx7\" (UniqueName: \"kubernetes.io/projected/a280b74a-cdbf-430b-9686-3055aaebf27e-kube-api-access-zstx7\") pod \"a280b74a-cdbf-430b-9686-3055aaebf27e\" (UID: \"a280b74a-cdbf-430b-9686-3055aaebf27e\") " Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.338759 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a280b74a-cdbf-430b-9686-3055aaebf27e-config\") pod \"a280b74a-cdbf-430b-9686-3055aaebf27e\" (UID: \"a280b74a-cdbf-430b-9686-3055aaebf27e\") " Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.338832 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a280b74a-cdbf-430b-9686-3055aaebf27e-dns-svc\") pod \"a280b74a-cdbf-430b-9686-3055aaebf27e\" (UID: \"a280b74a-cdbf-430b-9686-3055aaebf27e\") " Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.339296 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23-config" (OuterVolumeSpecName: "config") pod "26bf0d4f-2ed3-4428-8a62-4b3a1947fa23" (UID: "26bf0d4f-2ed3-4428-8a62-4b3a1947fa23"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.339850 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a280b74a-cdbf-430b-9686-3055aaebf27e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a280b74a-cdbf-430b-9686-3055aaebf27e" (UID: "a280b74a-cdbf-430b-9686-3055aaebf27e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.339972 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a280b74a-cdbf-430b-9686-3055aaebf27e-config" (OuterVolumeSpecName: "config") pod "a280b74a-cdbf-430b-9686-3055aaebf27e" (UID: "a280b74a-cdbf-430b-9686-3055aaebf27e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.347306 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dmpdv"] Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.347961 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a280b74a-cdbf-430b-9686-3055aaebf27e-kube-api-access-zstx7" (OuterVolumeSpecName: "kube-api-access-zstx7") pod "a280b74a-cdbf-430b-9686-3055aaebf27e" (UID: "a280b74a-cdbf-430b-9686-3055aaebf27e"). InnerVolumeSpecName "kube-api-access-zstx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.348414 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23-kube-api-access-km4x2" (OuterVolumeSpecName: "kube-api-access-km4x2") pod "26bf0d4f-2ed3-4428-8a62-4b3a1947fa23" (UID: "26bf0d4f-2ed3-4428-8a62-4b3a1947fa23"). InnerVolumeSpecName "kube-api-access-km4x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.437381 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gmp2c"] Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.442758 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.443200 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km4x2\" (UniqueName: \"kubernetes.io/projected/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23-kube-api-access-km4x2\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.443215 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zstx7\" (UniqueName: \"kubernetes.io/projected/a280b74a-cdbf-430b-9686-3055aaebf27e-kube-api-access-zstx7\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.443225 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a280b74a-cdbf-430b-9686-3055aaebf27e-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.443234 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a280b74a-cdbf-430b-9686-3055aaebf27e-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:04 crc kubenswrapper[4872]: W0126 09:22:04.462089 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod179dc633_3c59_4816_9ddf_db909211be5c.slice/crio-3c436cf1faea3cf029e0b07269b83da307159ee97f5b50fcea3f3082b628ea3a WatchSource:0}: Error finding container 3c436cf1faea3cf029e0b07269b83da307159ee97f5b50fcea3f3082b628ea3a: Status 404 returned error can't find the container with id 3c436cf1faea3cf029e0b07269b83da307159ee97f5b50fcea3f3082b628ea3a Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.530446 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.643594 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gmp2c" event={"ID":"cb72fdef-372c-41a1-bf57-6c71162ab194","Type":"ContainerStarted","Data":"53dd20c62a25f80dbd27a2c8fb52a0b84d20e9d700735bf0cce83c4040996719"} Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.648090 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" event={"ID":"26bf0d4f-2ed3-4428-8a62-4b3a1947fa23","Type":"ContainerDied","Data":"9b8556cdb52f6908c4d1ce583cdfbca555807483b1e8c1cb10dd9d6c724ee86b"} Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.648116 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-q4wxs" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.656375 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmpdv" event={"ID":"179dc633-3c59-4816-9ddf-db909211be5c","Type":"ContainerStarted","Data":"3c436cf1faea3cf029e0b07269b83da307159ee97f5b50fcea3f3082b628ea3a"} Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.658019 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" event={"ID":"a280b74a-cdbf-430b-9686-3055aaebf27e","Type":"ContainerDied","Data":"3554a7c53fc95890d863673f987d4669b5b5018a8f17ed026dcde0978ac63eae"} Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.658099 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-mfc4j" Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.662973 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bf0544f2-34dc-45fc-986e-cb2554c04566","Type":"ContainerStarted","Data":"6b5a48941d178d63636edd2c1e35c8f04c5bf2cb8f1e9e23bfe4180bc4cab7fd"} Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.666305 4872 generic.go:334] "Generic (PLEG): container finished" podID="e318d821-9c39-43ac-92e7-8834d9275c06" containerID="abb1541e39fb31702417e74775123fdee491af1be104c81929aaa5b53f3fbe16" exitCode=0 Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.666356 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkdtq" event={"ID":"e318d821-9c39-43ac-92e7-8834d9275c06","Type":"ContainerDied","Data":"abb1541e39fb31702417e74775123fdee491af1be104c81929aaa5b53f3fbe16"} Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.730940 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-q4wxs"] Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.736024 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-q4wxs"] Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.775289 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mfc4j"] Jan 26 09:22:04 crc kubenswrapper[4872]: I0126 09:22:04.789093 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-mfc4j"] Jan 26 09:22:05 crc kubenswrapper[4872]: I0126 09:22:05.201984 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26bf0d4f-2ed3-4428-8a62-4b3a1947fa23" path="/var/lib/kubelet/pods/26bf0d4f-2ed3-4428-8a62-4b3a1947fa23/volumes" Jan 26 09:22:05 crc kubenswrapper[4872]: I0126 09:22:05.202846 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a280b74a-cdbf-430b-9686-3055aaebf27e" path="/var/lib/kubelet/pods/a280b74a-cdbf-430b-9686-3055aaebf27e/volumes" Jan 26 09:22:05 crc kubenswrapper[4872]: I0126 09:22:05.242282 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Jan 26 09:22:05 crc kubenswrapper[4872]: I0126 09:22:05.390628 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mgn5r"] Jan 26 09:22:05 crc kubenswrapper[4872]: W0126 09:22:05.475665 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod204a2939_a385_4f7d_a08b_a008e6a14df0.slice/crio-fdcbfe06d2c0175d59d67ed70362b5b711f86b44eee342cc3308cf270d9eef45 WatchSource:0}: Error finding container fdcbfe06d2c0175d59d67ed70362b5b711f86b44eee342cc3308cf270d9eef45: Status 404 returned error can't find the container with id fdcbfe06d2c0175d59d67ed70362b5b711f86b44eee342cc3308cf270d9eef45 Jan 26 09:22:05 crc kubenswrapper[4872]: W0126 09:22:05.478487 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ece36ce_cbb6_4012_8b5c_87834502f26b.slice/crio-1e5bc09561bcef4aca37aa9a3be6b983f999cbdcd44c00e706f68a904ef9acfc WatchSource:0}: Error finding container 1e5bc09561bcef4aca37aa9a3be6b983f999cbdcd44c00e706f68a904ef9acfc: Status 404 returned error can't find the container with id 1e5bc09561bcef4aca37aa9a3be6b983f999cbdcd44c00e706f68a904ef9acfc Jan 26 09:22:05 crc kubenswrapper[4872]: I0126 09:22:05.678481 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mgn5r" event={"ID":"4ece36ce-cbb6-4012-8b5c-87834502f26b","Type":"ContainerStarted","Data":"1e5bc09561bcef4aca37aa9a3be6b983f999cbdcd44c00e706f68a904ef9acfc"} Jan 26 09:22:05 crc kubenswrapper[4872]: I0126 09:22:05.679841 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"204a2939-a385-4f7d-a08b-a008e6a14df0","Type":"ContainerStarted","Data":"fdcbfe06d2c0175d59d67ed70362b5b711f86b44eee342cc3308cf270d9eef45"} Jan 26 09:22:06 crc kubenswrapper[4872]: I0126 09:22:06.694094 4872 generic.go:334] "Generic (PLEG): container finished" podID="179dc633-3c59-4816-9ddf-db909211be5c" containerID="a609620847a579a7acc50c94ae51f6c0cbf975b7fa9df221e5737b8f68d81e89" exitCode=0 Jan 26 09:22:06 crc kubenswrapper[4872]: I0126 09:22:06.694605 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmpdv" event={"ID":"179dc633-3c59-4816-9ddf-db909211be5c","Type":"ContainerDied","Data":"a609620847a579a7acc50c94ae51f6c0cbf975b7fa9df221e5737b8f68d81e89"} Jan 26 09:22:06 crc kubenswrapper[4872]: I0126 09:22:06.697956 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3d16170b-8fc9-4c93-85fe-9ccef4da5bd0","Type":"ContainerStarted","Data":"bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696"} Jan 26 09:22:06 crc kubenswrapper[4872]: I0126 09:22:06.698460 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 26 09:22:06 crc kubenswrapper[4872]: I0126 09:22:06.701551 4872 generic.go:334] "Generic (PLEG): container finished" podID="e318d821-9c39-43ac-92e7-8834d9275c06" containerID="d26015abf0025ca498ba173fd4cf34744af428e3c1204b8dc059a805def8462d" exitCode=0 Jan 26 09:22:06 crc kubenswrapper[4872]: I0126 09:22:06.701620 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkdtq" event={"ID":"e318d821-9c39-43ac-92e7-8834d9275c06","Type":"ContainerDied","Data":"d26015abf0025ca498ba173fd4cf34744af428e3c1204b8dc059a805def8462d"} Jan 26 09:22:06 crc kubenswrapper[4872]: I0126 09:22:06.746214 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.880359374 podStartE2EDuration="27.746186297s" podCreationTimestamp="2026-01-26 09:21:39 +0000 UTC" firstStartedPulling="2026-01-26 09:21:40.66270935 +0000 UTC m=+833.971549151" lastFinishedPulling="2026-01-26 09:22:05.528536253 +0000 UTC m=+858.837376074" observedRunningTime="2026-01-26 09:22:06.736271128 +0000 UTC m=+860.045110929" watchObservedRunningTime="2026-01-26 09:22:06.746186297 +0000 UTC m=+860.055026098" Jan 26 09:22:09 crc kubenswrapper[4872]: I0126 09:22:09.738166 4872 generic.go:334] "Generic (PLEG): container finished" podID="179dc633-3c59-4816-9ddf-db909211be5c" containerID="501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592" exitCode=0 Jan 26 09:22:09 crc kubenswrapper[4872]: I0126 09:22:09.738842 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmpdv" event={"ID":"179dc633-3c59-4816-9ddf-db909211be5c","Type":"ContainerDied","Data":"501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592"} Jan 26 09:22:09 crc kubenswrapper[4872]: I0126 09:22:09.743831 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bf0544f2-34dc-45fc-986e-cb2554c04566","Type":"ContainerStarted","Data":"cdf8ded838ea887ec32f109d0cac88a4bf340a76d6bc90e051f2436c0213ab15"} Jan 26 09:22:09 crc kubenswrapper[4872]: I0126 09:22:09.754200 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkdtq" event={"ID":"e318d821-9c39-43ac-92e7-8834d9275c06","Type":"ContainerStarted","Data":"17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95"} Jan 26 09:22:09 crc kubenswrapper[4872]: I0126 09:22:09.762148 4872 generic.go:334] "Generic (PLEG): container finished" podID="4ece36ce-cbb6-4012-8b5c-87834502f26b" containerID="907272d6e20f1e712e8091e14b1ce7a7bff2bf73e441182b2c87d116c70406d5" exitCode=0 Jan 26 09:22:09 crc kubenswrapper[4872]: I0126 09:22:09.762262 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mgn5r" event={"ID":"4ece36ce-cbb6-4012-8b5c-87834502f26b","Type":"ContainerDied","Data":"907272d6e20f1e712e8091e14b1ce7a7bff2bf73e441182b2c87d116c70406d5"} Jan 26 09:22:09 crc kubenswrapper[4872]: I0126 09:22:09.764880 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gmp2c" event={"ID":"cb72fdef-372c-41a1-bf57-6c71162ab194","Type":"ContainerStarted","Data":"01dfd77cc39f1c0575b443a341b8359461cda5ea5707affee8a71adb1623d136"} Jan 26 09:22:09 crc kubenswrapper[4872]: I0126 09:22:09.765826 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-gmp2c" Jan 26 09:22:09 crc kubenswrapper[4872]: I0126 09:22:09.775537 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"204a2939-a385-4f7d-a08b-a008e6a14df0","Type":"ContainerStarted","Data":"d449dd03f8be10089657f46ffcea8e8fd92e7e362ea59cef8ca63534b3fd49e1"} Jan 26 09:22:09 crc kubenswrapper[4872]: I0126 09:22:09.793617 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bkdtq" podStartSLOduration=24.884054445 podStartE2EDuration="28.79358506s" podCreationTimestamp="2026-01-26 09:21:41 +0000 UTC" firstStartedPulling="2026-01-26 09:22:04.718692808 +0000 UTC m=+858.027532609" lastFinishedPulling="2026-01-26 09:22:08.628223423 +0000 UTC m=+861.937063224" observedRunningTime="2026-01-26 09:22:09.788837781 +0000 UTC m=+863.097677592" watchObservedRunningTime="2026-01-26 09:22:09.79358506 +0000 UTC m=+863.102424861" Jan 26 09:22:09 crc kubenswrapper[4872]: I0126 09:22:09.814649 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-gmp2c" podStartSLOduration=23.737404883 podStartE2EDuration="27.81461958s" podCreationTimestamp="2026-01-26 09:21:42 +0000 UTC" firstStartedPulling="2026-01-26 09:22:04.550703358 +0000 UTC m=+857.859543159" lastFinishedPulling="2026-01-26 09:22:08.627918055 +0000 UTC m=+861.936757856" observedRunningTime="2026-01-26 09:22:09.809643805 +0000 UTC m=+863.118483606" watchObservedRunningTime="2026-01-26 09:22:09.81461958 +0000 UTC m=+863.123459381" Jan 26 09:22:09 crc kubenswrapper[4872]: E0126 09:22:09.845214 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod179dc633_3c59_4816_9ddf_db909211be5c.slice/crio-conmon-501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod179dc633_3c59_4816_9ddf_db909211be5c.slice/crio-501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592.scope\": RecentStats: unable to find data in memory cache]" Jan 26 09:22:11 crc kubenswrapper[4872]: I0126 09:22:11.795532 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mgn5r" event={"ID":"4ece36ce-cbb6-4012-8b5c-87834502f26b","Type":"ContainerStarted","Data":"3ad4b50a265ea52ea2691be14cfdf1b2ccb0fb0462b9a04eb329a3a3e20c0a7a"} Jan 26 09:22:11 crc kubenswrapper[4872]: I0126 09:22:11.935999 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:22:11 crc kubenswrapper[4872]: I0126 09:22:11.936045 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:22:11 crc kubenswrapper[4872]: I0126 09:22:11.984664 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:22:12 crc kubenswrapper[4872]: I0126 09:22:12.815064 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmpdv" event={"ID":"179dc633-3c59-4816-9ddf-db909211be5c","Type":"ContainerStarted","Data":"7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06"} Jan 26 09:22:12 crc kubenswrapper[4872]: I0126 09:22:12.820329 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mgn5r" event={"ID":"4ece36ce-cbb6-4012-8b5c-87834502f26b","Type":"ContainerStarted","Data":"b732f716a199e8006e13392af42e8e8f69b050e5192df2445c384776a865dd29"} Jan 26 09:22:12 crc kubenswrapper[4872]: I0126 09:22:12.839447 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dmpdv" podStartSLOduration=15.419784325 podStartE2EDuration="19.839422128s" podCreationTimestamp="2026-01-26 09:21:53 +0000 UTC" firstStartedPulling="2026-01-26 09:22:07.65642516 +0000 UTC m=+860.965264971" lastFinishedPulling="2026-01-26 09:22:12.076062973 +0000 UTC m=+865.384902774" observedRunningTime="2026-01-26 09:22:12.837443787 +0000 UTC m=+866.146283598" watchObservedRunningTime="2026-01-26 09:22:12.839422128 +0000 UTC m=+866.148261939" Jan 26 09:22:12 crc kubenswrapper[4872]: I0126 09:22:12.866393 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-mgn5r" podStartSLOduration=27.75853393 podStartE2EDuration="30.866368266s" podCreationTimestamp="2026-01-26 09:21:42 +0000 UTC" firstStartedPulling="2026-01-26 09:22:05.519778352 +0000 UTC m=+858.828618163" lastFinishedPulling="2026-01-26 09:22:08.627612708 +0000 UTC m=+861.936452499" observedRunningTime="2026-01-26 09:22:12.860167859 +0000 UTC m=+866.169007660" watchObservedRunningTime="2026-01-26 09:22:12.866368266 +0000 UTC m=+866.175208077" Jan 26 09:22:13 crc kubenswrapper[4872]: I0126 09:22:13.219484 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:22:13 crc kubenswrapper[4872]: I0126 09:22:13.220004 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:22:14 crc kubenswrapper[4872]: I0126 09:22:14.240865 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:22:14 crc kubenswrapper[4872]: I0126 09:22:14.240922 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:22:15 crc kubenswrapper[4872]: I0126 09:22:15.305118 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dmpdv" podUID="179dc633-3c59-4816-9ddf-db909211be5c" containerName="registry-server" probeResult="failure" output=< Jan 26 09:22:15 crc kubenswrapper[4872]: timeout: failed to connect service ":50051" within 1s Jan 26 09:22:15 crc kubenswrapper[4872]: > Jan 26 09:22:19 crc kubenswrapper[4872]: I0126 09:22:19.859841 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 26 09:22:20 crc kubenswrapper[4872]: E0126 09:22:20.763584 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Jan 26 09:22:20 crc kubenswrapper[4872]: E0126 09:22:20.763822 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},EnvVar{Name:CONFIG_HASH,Value:nc4h98hcfh65fh59fhb4h674h58fh5dbhf5h567h647h545h76hbch5dbh5b6h645h547h695h677h7bh66fh68fhd4hbdh644hbh58bh567h5b8h59cq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c56wm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-nb-0_openstack(bf0544f2-34dc-45fc-986e-cb2554c04566): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:22:20 crc kubenswrapper[4872]: E0126 09:22:20.765416 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-nb-0" podUID="bf0544f2-34dc-45fc-986e-cb2554c04566" Jan 26 09:22:20 crc kubenswrapper[4872]: E0126 09:22:20.813674 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified" Jan 26 09:22:20 crc kubenswrapper[4872]: E0126 09:22:20.813957 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstack-network-exporter,Image:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,Command:[/app/openstack-network-exporter],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:OPENSTACK_NETWORK_EXPORTER_YAML,Value:/etc/config/openstack-network-exporter.yaml,ValueFrom:nil,},EnvVar{Name:CONFIG_HASH,Value:n577hbch65fh678h5fdh594h588hdch68dh5c6h657hcbh678h55dh5f9h5f8h97hcbh67bh68bh659h66bh6fh596h67h566h74h59ch5bdh555h548h59bq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovsdb-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovnmetrics.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovnmetrics.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-certs-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4ssz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovsdbserver-sb-0_openstack(204a2939-a385-4f7d-a08b-a008e6a14df0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:22:20 crc kubenswrapper[4872]: E0126 09:22:20.815669 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovsdbserver-sb-0" podUID="204a2939-a385-4f7d-a08b-a008e6a14df0" Jan 26 09:22:20 crc kubenswrapper[4872]: E0126 09:22:20.924935 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bf0544f2-34dc-45fc-986e-cb2554c04566" Jan 26 09:22:20 crc kubenswrapper[4872]: E0126 09:22:20.926638 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="204a2939-a385-4f7d-a08b-a008e6a14df0" Jan 26 09:22:21 crc kubenswrapper[4872]: I0126 09:22:21.895176 4872 generic.go:334] "Generic (PLEG): container finished" podID="130ac93b-f4db-4e8b-9637-f05cda3d1f0c" containerID="ebf6afebed792a45b43a70bca753d1fdb5f85a3773d2b1bf63dfb0edeb146680" exitCode=0 Jan 26 09:22:21 crc kubenswrapper[4872]: I0126 09:22:21.895250 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" event={"ID":"130ac93b-f4db-4e8b-9637-f05cda3d1f0c","Type":"ContainerDied","Data":"ebf6afebed792a45b43a70bca753d1fdb5f85a3773d2b1bf63dfb0edeb146680"} Jan 26 09:22:21 crc kubenswrapper[4872]: I0126 09:22:21.898627 4872 generic.go:334] "Generic (PLEG): container finished" podID="35971ebc-86cc-4a97-b3ba-ebe63c6f1200" containerID="358efcc88abfe41bea21be886fc97f03f4bf7145ce74f3540c858f2facdcb644" exitCode=0 Jan 26 09:22:21 crc kubenswrapper[4872]: I0126 09:22:21.898689 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" event={"ID":"35971ebc-86cc-4a97-b3ba-ebe63c6f1200","Type":"ContainerDied","Data":"358efcc88abfe41bea21be886fc97f03f4bf7145ce74f3540c858f2facdcb644"} Jan 26 09:22:21 crc kubenswrapper[4872]: I0126 09:22:21.905359 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ab5ce366-26e9-424a-9a73-9173636c2902","Type":"ContainerStarted","Data":"32fb13a4904eea9a07ba101c18051724d1dceef52b02f18cca27b27d1312c400"} Jan 26 09:22:21 crc kubenswrapper[4872]: I0126 09:22:21.909335 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"98a9098c-4abc-4a20-be26-d62bd842bd04","Type":"ContainerStarted","Data":"fcf7d6b28ecbdd705e8543e1d2ff844a948f043073336a9cae4de6aed2b1701e"} Jan 26 09:22:21 crc kubenswrapper[4872]: I0126 09:22:21.909576 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Jan 26 09:22:21 crc kubenswrapper[4872]: I0126 09:22:21.911964 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9082961a-1e93-49b5-8420-7516b675ea6f","Type":"ContainerStarted","Data":"15d0d5675c14184152e24b1c7b918cb06938ca91c71d76fd99a2a99a1182c168"} Jan 26 09:22:22 crc kubenswrapper[4872]: I0126 09:22:22.007053 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.232122713 podStartE2EDuration="45.007030488s" podCreationTimestamp="2026-01-26 09:21:37 +0000 UTC" firstStartedPulling="2026-01-26 09:21:38.157960042 +0000 UTC m=+831.466799843" lastFinishedPulling="2026-01-26 09:22:20.932867827 +0000 UTC m=+874.241707618" observedRunningTime="2026-01-26 09:22:21.99360515 +0000 UTC m=+875.302444961" watchObservedRunningTime="2026-01-26 09:22:22.007030488 +0000 UTC m=+875.315870289" Jan 26 09:22:22 crc kubenswrapper[4872]: I0126 09:22:22.009466 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:22:22 crc kubenswrapper[4872]: I0126 09:22:22.089499 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Jan 26 09:22:22 crc kubenswrapper[4872]: E0126 09:22:22.093088 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="204a2939-a385-4f7d-a08b-a008e6a14df0" Jan 26 09:22:22 crc kubenswrapper[4872]: I0126 09:22:22.094132 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bkdtq"] Jan 26 09:22:22 crc kubenswrapper[4872]: I0126 09:22:22.921352 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0792ad43-38af-4c94-bd3b-fc7d3a7af5da","Type":"ContainerStarted","Data":"61e80e60ff53238778a51ec238b504652ca6a6c35b0dc148831642ff362f238a"} Jan 26 09:22:22 crc kubenswrapper[4872]: I0126 09:22:22.923503 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" event={"ID":"130ac93b-f4db-4e8b-9637-f05cda3d1f0c","Type":"ContainerStarted","Data":"91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec"} Jan 26 09:22:22 crc kubenswrapper[4872]: I0126 09:22:22.924183 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:22:22 crc kubenswrapper[4872]: I0126 09:22:22.926566 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" event={"ID":"35971ebc-86cc-4a97-b3ba-ebe63c6f1200","Type":"ContainerStarted","Data":"724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e"} Jan 26 09:22:22 crc kubenswrapper[4872]: I0126 09:22:22.927712 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:22:22 crc kubenswrapper[4872]: I0126 09:22:22.928175 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"82c8133d-d299-4c99-b8a0-96099e834e8c","Type":"ContainerStarted","Data":"ad7fd4226f3c6d842b398d7962b8022bc46582f511d8ebd06feb9754c684ed0d"} Jan 26 09:22:22 crc kubenswrapper[4872]: I0126 09:22:22.928912 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bkdtq" podUID="e318d821-9c39-43ac-92e7-8834d9275c06" containerName="registry-server" containerID="cri-o://17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95" gracePeriod=2 Jan 26 09:22:22 crc kubenswrapper[4872]: I0126 09:22:22.994237 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" podStartSLOduration=3.02235721 podStartE2EDuration="49.994201568s" podCreationTimestamp="2026-01-26 09:21:33 +0000 UTC" firstStartedPulling="2026-01-26 09:21:34.014412753 +0000 UTC m=+827.323252554" lastFinishedPulling="2026-01-26 09:22:20.986257111 +0000 UTC m=+874.295096912" observedRunningTime="2026-01-26 09:22:22.987206752 +0000 UTC m=+876.296046573" watchObservedRunningTime="2026-01-26 09:22:22.994201568 +0000 UTC m=+876.303041369" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.011873 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" podStartSLOduration=4.132121595 podStartE2EDuration="51.011846273s" podCreationTimestamp="2026-01-26 09:21:32 +0000 UTC" firstStartedPulling="2026-01-26 09:21:34.077988775 +0000 UTC m=+827.386828576" lastFinishedPulling="2026-01-26 09:22:20.957713453 +0000 UTC m=+874.266553254" observedRunningTime="2026-01-26 09:22:23.010646432 +0000 UTC m=+876.319486233" watchObservedRunningTime="2026-01-26 09:22:23.011846273 +0000 UTC m=+876.320686074" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.058156 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Jan 26 09:22:23 crc kubenswrapper[4872]: E0126 09:22:23.060385 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bf0544f2-34dc-45fc-986e-cb2554c04566" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.091150 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Jan 26 09:22:23 crc kubenswrapper[4872]: E0126 09:22:23.093633 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="204a2939-a385-4f7d-a08b-a008e6a14df0" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.112564 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.153913 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.414511 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.542395 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e318d821-9c39-43ac-92e7-8834d9275c06-utilities\") pod \"e318d821-9c39-43ac-92e7-8834d9275c06\" (UID: \"e318d821-9c39-43ac-92e7-8834d9275c06\") " Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.542622 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e318d821-9c39-43ac-92e7-8834d9275c06-catalog-content\") pod \"e318d821-9c39-43ac-92e7-8834d9275c06\" (UID: \"e318d821-9c39-43ac-92e7-8834d9275c06\") " Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.543344 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e318d821-9c39-43ac-92e7-8834d9275c06-utilities" (OuterVolumeSpecName: "utilities") pod "e318d821-9c39-43ac-92e7-8834d9275c06" (UID: "e318d821-9c39-43ac-92e7-8834d9275c06"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.545435 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh6bp\" (UniqueName: \"kubernetes.io/projected/e318d821-9c39-43ac-92e7-8834d9275c06-kube-api-access-hh6bp\") pod \"e318d821-9c39-43ac-92e7-8834d9275c06\" (UID: \"e318d821-9c39-43ac-92e7-8834d9275c06\") " Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.546380 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e318d821-9c39-43ac-92e7-8834d9275c06-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.553041 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e318d821-9c39-43ac-92e7-8834d9275c06-kube-api-access-hh6bp" (OuterVolumeSpecName: "kube-api-access-hh6bp") pod "e318d821-9c39-43ac-92e7-8834d9275c06" (UID: "e318d821-9c39-43ac-92e7-8834d9275c06"). InnerVolumeSpecName "kube-api-access-hh6bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.597349 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e318d821-9c39-43ac-92e7-8834d9275c06-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e318d821-9c39-43ac-92e7-8834d9275c06" (UID: "e318d821-9c39-43ac-92e7-8834d9275c06"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.648721 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e318d821-9c39-43ac-92e7-8834d9275c06-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.648770 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh6bp\" (UniqueName: \"kubernetes.io/projected/e318d821-9c39-43ac-92e7-8834d9275c06-kube-api-access-hh6bp\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.938845 4872 generic.go:334] "Generic (PLEG): container finished" podID="e318d821-9c39-43ac-92e7-8834d9275c06" containerID="17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95" exitCode=0 Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.938941 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bkdtq" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.938954 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkdtq" event={"ID":"e318d821-9c39-43ac-92e7-8834d9275c06","Type":"ContainerDied","Data":"17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95"} Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.939641 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bkdtq" event={"ID":"e318d821-9c39-43ac-92e7-8834d9275c06","Type":"ContainerDied","Data":"e9d7dc247b33851f470f99549e1670016adcb90c412e43b62326b06e8cb7047f"} Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.939674 4872 scope.go:117] "RemoveContainer" containerID="17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.940380 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Jan 26 09:22:23 crc kubenswrapper[4872]: E0126 09:22:23.945973 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bf0544f2-34dc-45fc-986e-cb2554c04566" Jan 26 09:22:23 crc kubenswrapper[4872]: E0126 09:22:23.956527 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="204a2939-a385-4f7d-a08b-a008e6a14df0" Jan 26 09:22:23 crc kubenswrapper[4872]: I0126 09:22:23.987031 4872 scope.go:117] "RemoveContainer" containerID="d26015abf0025ca498ba173fd4cf34744af428e3c1204b8dc059a805def8462d" Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.003495 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.007703 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bkdtq"] Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.025444 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bkdtq"] Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.041214 4872 scope.go:117] "RemoveContainer" containerID="abb1541e39fb31702417e74775123fdee491af1be104c81929aaa5b53f3fbe16" Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.041402 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.075284 4872 scope.go:117] "RemoveContainer" containerID="17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95" Jan 26 09:22:24 crc kubenswrapper[4872]: E0126 09:22:24.076935 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95\": container with ID starting with 17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95 not found: ID does not exist" containerID="17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95" Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.076994 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95"} err="failed to get container status \"17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95\": rpc error: code = NotFound desc = could not find container \"17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95\": container with ID starting with 17ac01230b7e18c9f09841eb4e6131db667b400b9fe05622e22d56deacc44f95 not found: ID does not exist" Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.077020 4872 scope.go:117] "RemoveContainer" containerID="d26015abf0025ca498ba173fd4cf34744af428e3c1204b8dc059a805def8462d" Jan 26 09:22:24 crc kubenswrapper[4872]: E0126 09:22:24.079041 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d26015abf0025ca498ba173fd4cf34744af428e3c1204b8dc059a805def8462d\": container with ID starting with d26015abf0025ca498ba173fd4cf34744af428e3c1204b8dc059a805def8462d not found: ID does not exist" containerID="d26015abf0025ca498ba173fd4cf34744af428e3c1204b8dc059a805def8462d" Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.079078 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d26015abf0025ca498ba173fd4cf34744af428e3c1204b8dc059a805def8462d"} err="failed to get container status \"d26015abf0025ca498ba173fd4cf34744af428e3c1204b8dc059a805def8462d\": rpc error: code = NotFound desc = could not find container \"d26015abf0025ca498ba173fd4cf34744af428e3c1204b8dc059a805def8462d\": container with ID starting with d26015abf0025ca498ba173fd4cf34744af428e3c1204b8dc059a805def8462d not found: ID does not exist" Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.079311 4872 scope.go:117] "RemoveContainer" containerID="abb1541e39fb31702417e74775123fdee491af1be104c81929aaa5b53f3fbe16" Jan 26 09:22:24 crc kubenswrapper[4872]: E0126 09:22:24.080154 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abb1541e39fb31702417e74775123fdee491af1be104c81929aaa5b53f3fbe16\": container with ID starting with abb1541e39fb31702417e74775123fdee491af1be104c81929aaa5b53f3fbe16 not found: ID does not exist" containerID="abb1541e39fb31702417e74775123fdee491af1be104c81929aaa5b53f3fbe16" Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.080315 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abb1541e39fb31702417e74775123fdee491af1be104c81929aaa5b53f3fbe16"} err="failed to get container status \"abb1541e39fb31702417e74775123fdee491af1be104c81929aaa5b53f3fbe16\": rpc error: code = NotFound desc = could not find container \"abb1541e39fb31702417e74775123fdee491af1be104c81929aaa5b53f3fbe16\": container with ID starting with abb1541e39fb31702417e74775123fdee491af1be104c81929aaa5b53f3fbe16 not found: ID does not exist" Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.301825 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:22:24 crc kubenswrapper[4872]: I0126 09:22:24.343637 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:22:24 crc kubenswrapper[4872]: E0126 09:22:24.953429 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-sb-0" podUID="204a2939-a385-4f7d-a08b-a008e6a14df0" Jan 26 09:22:24 crc kubenswrapper[4872]: E0126 09:22:24.953549 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bf0544f2-34dc-45fc-986e-cb2554c04566" Jan 26 09:22:25 crc kubenswrapper[4872]: I0126 09:22:25.195938 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e318d821-9c39-43ac-92e7-8834d9275c06" path="/var/lib/kubelet/pods/e318d821-9c39-43ac-92e7-8834d9275c06/volumes" Jan 26 09:22:25 crc kubenswrapper[4872]: I0126 09:22:25.260024 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dmpdv"] Jan 26 09:22:25 crc kubenswrapper[4872]: I0126 09:22:25.960746 4872 generic.go:334] "Generic (PLEG): container finished" podID="9082961a-1e93-49b5-8420-7516b675ea6f" containerID="15d0d5675c14184152e24b1c7b918cb06938ca91c71d76fd99a2a99a1182c168" exitCode=0 Jan 26 09:22:25 crc kubenswrapper[4872]: I0126 09:22:25.960841 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9082961a-1e93-49b5-8420-7516b675ea6f","Type":"ContainerDied","Data":"15d0d5675c14184152e24b1c7b918cb06938ca91c71d76fd99a2a99a1182c168"} Jan 26 09:22:25 crc kubenswrapper[4872]: I0126 09:22:25.961022 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dmpdv" podUID="179dc633-3c59-4816-9ddf-db909211be5c" containerName="registry-server" containerID="cri-o://7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06" gracePeriod=2 Jan 26 09:22:25 crc kubenswrapper[4872]: E0126 09:22:25.962656 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstack-network-exporter\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified\\\"\"" pod="openstack/ovsdbserver-nb-0" podUID="bf0544f2-34dc-45fc-986e-cb2554c04566" Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.356145 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.501823 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7wfd\" (UniqueName: \"kubernetes.io/projected/179dc633-3c59-4816-9ddf-db909211be5c-kube-api-access-k7wfd\") pod \"179dc633-3c59-4816-9ddf-db909211be5c\" (UID: \"179dc633-3c59-4816-9ddf-db909211be5c\") " Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.501962 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/179dc633-3c59-4816-9ddf-db909211be5c-utilities\") pod \"179dc633-3c59-4816-9ddf-db909211be5c\" (UID: \"179dc633-3c59-4816-9ddf-db909211be5c\") " Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.502026 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/179dc633-3c59-4816-9ddf-db909211be5c-catalog-content\") pod \"179dc633-3c59-4816-9ddf-db909211be5c\" (UID: \"179dc633-3c59-4816-9ddf-db909211be5c\") " Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.503615 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/179dc633-3c59-4816-9ddf-db909211be5c-utilities" (OuterVolumeSpecName: "utilities") pod "179dc633-3c59-4816-9ddf-db909211be5c" (UID: "179dc633-3c59-4816-9ddf-db909211be5c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.508032 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/179dc633-3c59-4816-9ddf-db909211be5c-kube-api-access-k7wfd" (OuterVolumeSpecName: "kube-api-access-k7wfd") pod "179dc633-3c59-4816-9ddf-db909211be5c" (UID: "179dc633-3c59-4816-9ddf-db909211be5c"). InnerVolumeSpecName "kube-api-access-k7wfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.605119 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7wfd\" (UniqueName: \"kubernetes.io/projected/179dc633-3c59-4816-9ddf-db909211be5c-kube-api-access-k7wfd\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.605305 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/179dc633-3c59-4816-9ddf-db909211be5c-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.625703 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/179dc633-3c59-4816-9ddf-db909211be5c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "179dc633-3c59-4816-9ddf-db909211be5c" (UID: "179dc633-3c59-4816-9ddf-db909211be5c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.707366 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/179dc633-3c59-4816-9ddf-db909211be5c-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.971490 4872 generic.go:334] "Generic (PLEG): container finished" podID="ab5ce366-26e9-424a-9a73-9173636c2902" containerID="32fb13a4904eea9a07ba101c18051724d1dceef52b02f18cca27b27d1312c400" exitCode=0 Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.971577 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ab5ce366-26e9-424a-9a73-9173636c2902","Type":"ContainerDied","Data":"32fb13a4904eea9a07ba101c18051724d1dceef52b02f18cca27b27d1312c400"} Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.973907 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9082961a-1e93-49b5-8420-7516b675ea6f","Type":"ContainerStarted","Data":"13a67f1db1e86932bee65051046d5c5b291dd9dcb2150c4c07050b05247fb3f0"} Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.978040 4872 generic.go:334] "Generic (PLEG): container finished" podID="179dc633-3c59-4816-9ddf-db909211be5c" containerID="7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06" exitCode=0 Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.978096 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmpdv" event={"ID":"179dc633-3c59-4816-9ddf-db909211be5c","Type":"ContainerDied","Data":"7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06"} Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.978136 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dmpdv" event={"ID":"179dc633-3c59-4816-9ddf-db909211be5c","Type":"ContainerDied","Data":"3c436cf1faea3cf029e0b07269b83da307159ee97f5b50fcea3f3082b628ea3a"} Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.978159 4872 scope.go:117] "RemoveContainer" containerID="7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06" Jan 26 09:22:26 crc kubenswrapper[4872]: I0126 09:22:26.978317 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dmpdv" Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.051019 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.908106806 podStartE2EDuration="53.050991252s" podCreationTimestamp="2026-01-26 09:21:34 +0000 UTC" firstStartedPulling="2026-01-26 09:21:36.842471223 +0000 UTC m=+830.151311014" lastFinishedPulling="2026-01-26 09:22:20.985355659 +0000 UTC m=+874.294195460" observedRunningTime="2026-01-26 09:22:27.037613434 +0000 UTC m=+880.346453235" watchObservedRunningTime="2026-01-26 09:22:27.050991252 +0000 UTC m=+880.359831063" Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.079669 4872 scope.go:117] "RemoveContainer" containerID="501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592" Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.105523 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dmpdv"] Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.112842 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dmpdv"] Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.151076 4872 scope.go:117] "RemoveContainer" containerID="a609620847a579a7acc50c94ae51f6c0cbf975b7fa9df221e5737b8f68d81e89" Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.207291 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="179dc633-3c59-4816-9ddf-db909211be5c" path="/var/lib/kubelet/pods/179dc633-3c59-4816-9ddf-db909211be5c/volumes" Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.220138 4872 scope.go:117] "RemoveContainer" containerID="7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06" Jan 26 09:22:27 crc kubenswrapper[4872]: E0126 09:22:27.221056 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06\": container with ID starting with 7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06 not found: ID does not exist" containerID="7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06" Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.221127 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06"} err="failed to get container status \"7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06\": rpc error: code = NotFound desc = could not find container \"7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06\": container with ID starting with 7b317b59baa5251b50c0109ce855984f7d305763e77cc57e0db35c33ad174d06 not found: ID does not exist" Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.221169 4872 scope.go:117] "RemoveContainer" containerID="501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592" Jan 26 09:22:27 crc kubenswrapper[4872]: E0126 09:22:27.221986 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592\": container with ID starting with 501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592 not found: ID does not exist" containerID="501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592" Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.222034 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592"} err="failed to get container status \"501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592\": rpc error: code = NotFound desc = could not find container \"501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592\": container with ID starting with 501e818ddbe0cbb914488c26a3292b64b37da97b9f2481544edd2a51effad592 not found: ID does not exist" Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.222052 4872 scope.go:117] "RemoveContainer" containerID="a609620847a579a7acc50c94ae51f6c0cbf975b7fa9df221e5737b8f68d81e89" Jan 26 09:22:27 crc kubenswrapper[4872]: E0126 09:22:27.226951 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a609620847a579a7acc50c94ae51f6c0cbf975b7fa9df221e5737b8f68d81e89\": container with ID starting with a609620847a579a7acc50c94ae51f6c0cbf975b7fa9df221e5737b8f68d81e89 not found: ID does not exist" containerID="a609620847a579a7acc50c94ae51f6c0cbf975b7fa9df221e5737b8f68d81e89" Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.227012 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a609620847a579a7acc50c94ae51f6c0cbf975b7fa9df221e5737b8f68d81e89"} err="failed to get container status \"a609620847a579a7acc50c94ae51f6c0cbf975b7fa9df221e5737b8f68d81e89\": rpc error: code = NotFound desc = could not find container \"a609620847a579a7acc50c94ae51f6c0cbf975b7fa9df221e5737b8f68d81e89\": container with ID starting with a609620847a579a7acc50c94ae51f6c0cbf975b7fa9df221e5737b8f68d81e89 not found: ID does not exist" Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.449751 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Jan 26 09:22:27 crc kubenswrapper[4872]: I0126 09:22:27.993020 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"ab5ce366-26e9-424a-9a73-9173636c2902","Type":"ContainerStarted","Data":"2f955b31fb8d30de2e62f4e75b3e8f7c809f6fb0cde662dae7055e7defaba0d2"} Jan 26 09:22:28 crc kubenswrapper[4872]: I0126 09:22:28.029951 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=10.686280448 podStartE2EDuration="53.029926574s" podCreationTimestamp="2026-01-26 09:21:35 +0000 UTC" firstStartedPulling="2026-01-26 09:21:38.503303478 +0000 UTC m=+831.812143269" lastFinishedPulling="2026-01-26 09:22:20.846949584 +0000 UTC m=+874.155789395" observedRunningTime="2026-01-26 09:22:28.025032701 +0000 UTC m=+881.333872522" watchObservedRunningTime="2026-01-26 09:22:28.029926574 +0000 UTC m=+881.338766375" Jan 26 09:22:28 crc kubenswrapper[4872]: I0126 09:22:28.087140 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:22:28 crc kubenswrapper[4872]: I0126 09:22:28.414016 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:22:28 crc kubenswrapper[4872]: I0126 09:22:28.469621 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cb2wp"] Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.002857 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" podUID="35971ebc-86cc-4a97-b3ba-ebe63c6f1200" containerName="dnsmasq-dns" containerID="cri-o://724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e" gracePeriod=10 Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.503264 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-rm64t"] Jan 26 09:22:29 crc kubenswrapper[4872]: E0126 09:22:29.508882 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c69d55a-db50-4e0a-a86e-5b92f8088020" containerName="registry-server" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.508933 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c69d55a-db50-4e0a-a86e-5b92f8088020" containerName="registry-server" Jan 26 09:22:29 crc kubenswrapper[4872]: E0126 09:22:29.508964 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="179dc633-3c59-4816-9ddf-db909211be5c" containerName="registry-server" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.508970 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="179dc633-3c59-4816-9ddf-db909211be5c" containerName="registry-server" Jan 26 09:22:29 crc kubenswrapper[4872]: E0126 09:22:29.508992 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c69d55a-db50-4e0a-a86e-5b92f8088020" containerName="extract-utilities" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.508999 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c69d55a-db50-4e0a-a86e-5b92f8088020" containerName="extract-utilities" Jan 26 09:22:29 crc kubenswrapper[4872]: E0126 09:22:29.509012 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c69d55a-db50-4e0a-a86e-5b92f8088020" containerName="extract-content" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.509018 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c69d55a-db50-4e0a-a86e-5b92f8088020" containerName="extract-content" Jan 26 09:22:29 crc kubenswrapper[4872]: E0126 09:22:29.509029 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e318d821-9c39-43ac-92e7-8834d9275c06" containerName="extract-content" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.509035 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e318d821-9c39-43ac-92e7-8834d9275c06" containerName="extract-content" Jan 26 09:22:29 crc kubenswrapper[4872]: E0126 09:22:29.509044 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="179dc633-3c59-4816-9ddf-db909211be5c" containerName="extract-utilities" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.509050 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="179dc633-3c59-4816-9ddf-db909211be5c" containerName="extract-utilities" Jan 26 09:22:29 crc kubenswrapper[4872]: E0126 09:22:29.509060 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="179dc633-3c59-4816-9ddf-db909211be5c" containerName="extract-content" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.509066 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="179dc633-3c59-4816-9ddf-db909211be5c" containerName="extract-content" Jan 26 09:22:29 crc kubenswrapper[4872]: E0126 09:22:29.509081 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e318d821-9c39-43ac-92e7-8834d9275c06" containerName="extract-utilities" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.509086 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e318d821-9c39-43ac-92e7-8834d9275c06" containerName="extract-utilities" Jan 26 09:22:29 crc kubenswrapper[4872]: E0126 09:22:29.509098 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e318d821-9c39-43ac-92e7-8834d9275c06" containerName="registry-server" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.509103 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e318d821-9c39-43ac-92e7-8834d9275c06" containerName="registry-server" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.509252 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="179dc633-3c59-4816-9ddf-db909211be5c" containerName="registry-server" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.509276 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c69d55a-db50-4e0a-a86e-5b92f8088020" containerName="registry-server" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.509283 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e318d821-9c39-43ac-92e7-8834d9275c06" containerName="registry-server" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.510263 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.537443 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-rm64t"] Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.569582 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b46a9b4-7d84-46a7-82f2-6093f2563f51-config\") pod \"dnsmasq-dns-7cb5889db5-rm64t\" (UID: \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\") " pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.570305 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b46a9b4-7d84-46a7-82f2-6093f2563f51-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-rm64t\" (UID: \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\") " pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.570359 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hskl\" (UniqueName: \"kubernetes.io/projected/2b46a9b4-7d84-46a7-82f2-6093f2563f51-kube-api-access-6hskl\") pod \"dnsmasq-dns-7cb5889db5-rm64t\" (UID: \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\") " pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.636138 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.672050 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b46a9b4-7d84-46a7-82f2-6093f2563f51-config\") pod \"dnsmasq-dns-7cb5889db5-rm64t\" (UID: \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\") " pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.672180 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b46a9b4-7d84-46a7-82f2-6093f2563f51-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-rm64t\" (UID: \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\") " pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.672222 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hskl\" (UniqueName: \"kubernetes.io/projected/2b46a9b4-7d84-46a7-82f2-6093f2563f51-kube-api-access-6hskl\") pod \"dnsmasq-dns-7cb5889db5-rm64t\" (UID: \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\") " pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.673752 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b46a9b4-7d84-46a7-82f2-6093f2563f51-config\") pod \"dnsmasq-dns-7cb5889db5-rm64t\" (UID: \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\") " pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.682067 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b46a9b4-7d84-46a7-82f2-6093f2563f51-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-rm64t\" (UID: \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\") " pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.748957 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hskl\" (UniqueName: \"kubernetes.io/projected/2b46a9b4-7d84-46a7-82f2-6093f2563f51-kube-api-access-6hskl\") pod \"dnsmasq-dns-7cb5889db5-rm64t\" (UID: \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\") " pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.774052 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6f254\" (UniqueName: \"kubernetes.io/projected/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-kube-api-access-6f254\") pod \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\" (UID: \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\") " Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.774254 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-dns-svc\") pod \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\" (UID: \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\") " Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.774382 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-config\") pod \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\" (UID: \"35971ebc-86cc-4a97-b3ba-ebe63c6f1200\") " Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.787937 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-kube-api-access-6f254" (OuterVolumeSpecName: "kube-api-access-6f254") pod "35971ebc-86cc-4a97-b3ba-ebe63c6f1200" (UID: "35971ebc-86cc-4a97-b3ba-ebe63c6f1200"). InnerVolumeSpecName "kube-api-access-6f254". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.819905 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-config" (OuterVolumeSpecName: "config") pod "35971ebc-86cc-4a97-b3ba-ebe63c6f1200" (UID: "35971ebc-86cc-4a97-b3ba-ebe63c6f1200"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.840940 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "35971ebc-86cc-4a97-b3ba-ebe63c6f1200" (UID: "35971ebc-86cc-4a97-b3ba-ebe63c6f1200"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.850406 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.876920 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6f254\" (UniqueName: \"kubernetes.io/projected/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-kube-api-access-6f254\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.876968 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:29 crc kubenswrapper[4872]: I0126 09:22:29.876982 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35971ebc-86cc-4a97-b3ba-ebe63c6f1200-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.032347 4872 generic.go:334] "Generic (PLEG): container finished" podID="35971ebc-86cc-4a97-b3ba-ebe63c6f1200" containerID="724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e" exitCode=0 Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.032404 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" event={"ID":"35971ebc-86cc-4a97-b3ba-ebe63c6f1200","Type":"ContainerDied","Data":"724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e"} Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.032444 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" event={"ID":"35971ebc-86cc-4a97-b3ba-ebe63c6f1200","Type":"ContainerDied","Data":"092c4fd9d43f568d59845a515ef487ab4fe967d46d850b6b56cd1d948366a5a7"} Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.032469 4872 scope.go:117] "RemoveContainer" containerID="724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.032626 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-cb2wp" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.094449 4872 scope.go:117] "RemoveContainer" containerID="358efcc88abfe41bea21be886fc97f03f4bf7145ce74f3540c858f2facdcb644" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.200496 4872 scope.go:117] "RemoveContainer" containerID="724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e" Jan 26 09:22:30 crc kubenswrapper[4872]: E0126 09:22:30.201299 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e\": container with ID starting with 724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e not found: ID does not exist" containerID="724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.203234 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e"} err="failed to get container status \"724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e\": rpc error: code = NotFound desc = could not find container \"724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e\": container with ID starting with 724acdbedc2aed63240d6d9b25d009edef309c9a010b56dc7a1f3d591d4be09e not found: ID does not exist" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.203432 4872 scope.go:117] "RemoveContainer" containerID="358efcc88abfe41bea21be886fc97f03f4bf7145ce74f3540c858f2facdcb644" Jan 26 09:22:30 crc kubenswrapper[4872]: E0126 09:22:30.209371 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"358efcc88abfe41bea21be886fc97f03f4bf7145ce74f3540c858f2facdcb644\": container with ID starting with 358efcc88abfe41bea21be886fc97f03f4bf7145ce74f3540c858f2facdcb644 not found: ID does not exist" containerID="358efcc88abfe41bea21be886fc97f03f4bf7145ce74f3540c858f2facdcb644" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.209779 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"358efcc88abfe41bea21be886fc97f03f4bf7145ce74f3540c858f2facdcb644"} err="failed to get container status \"358efcc88abfe41bea21be886fc97f03f4bf7145ce74f3540c858f2facdcb644\": rpc error: code = NotFound desc = could not find container \"358efcc88abfe41bea21be886fc97f03f4bf7145ce74f3540c858f2facdcb644\": container with ID starting with 358efcc88abfe41bea21be886fc97f03f4bf7145ce74f3540c858f2facdcb644 not found: ID does not exist" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.237082 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cb2wp"] Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.271564 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-cb2wp"] Jan 26 09:22:30 crc kubenswrapper[4872]: E0126 09:22:30.294121 4872 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.70:58242->38.102.83.70:38695: read tcp 38.102.83.70:58242->38.102.83.70:38695: read: connection reset by peer Jan 26 09:22:30 crc kubenswrapper[4872]: E0126 09:22:30.294658 4872 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.70:58242->38.102.83.70:38695: write tcp 38.102.83.70:58242->38.102.83.70:38695: write: broken pipe Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.517903 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-rm64t"] Jan 26 09:22:30 crc kubenswrapper[4872]: W0126 09:22:30.526590 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b46a9b4_7d84_46a7_82f2_6093f2563f51.slice/crio-8a35796980197b34a5c4bca0dd090c652c3902052ec3925930778738e1dfc28c WatchSource:0}: Error finding container 8a35796980197b34a5c4bca0dd090c652c3902052ec3925930778738e1dfc28c: Status 404 returned error can't find the container with id 8a35796980197b34a5c4bca0dd090c652c3902052ec3925930778738e1dfc28c Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.675911 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Jan 26 09:22:30 crc kubenswrapper[4872]: E0126 09:22:30.676336 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35971ebc-86cc-4a97-b3ba-ebe63c6f1200" containerName="dnsmasq-dns" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.676523 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="35971ebc-86cc-4a97-b3ba-ebe63c6f1200" containerName="dnsmasq-dns" Jan 26 09:22:30 crc kubenswrapper[4872]: E0126 09:22:30.676585 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35971ebc-86cc-4a97-b3ba-ebe63c6f1200" containerName="init" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.676600 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="35971ebc-86cc-4a97-b3ba-ebe63c6f1200" containerName="init" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.677000 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="35971ebc-86cc-4a97-b3ba-ebe63c6f1200" containerName="dnsmasq-dns" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.682701 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.687706 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.687761 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-knxbl" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.688048 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.689488 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.702414 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.707356 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.707531 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-cache\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.707618 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v8qj\" (UniqueName: \"kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-kube-api-access-8v8qj\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.707748 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-lock\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.707934 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.708095 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.810041 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-cache\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.810130 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v8qj\" (UniqueName: \"kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-kube-api-access-8v8qj\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.810249 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-lock\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.810349 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.810490 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.810543 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.810591 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-cache\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.810723 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-lock\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: E0126 09:22:30.810888 4872 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 26 09:22:30 crc kubenswrapper[4872]: E0126 09:22:30.810917 4872 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 26 09:22:30 crc kubenswrapper[4872]: E0126 09:22:30.810975 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift podName:f94cf23a-424d-4bd6-8e19-b1f3dc7a0160 nodeName:}" failed. No retries permitted until 2026-01-26 09:22:31.310952879 +0000 UTC m=+884.619792680 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift") pod "swift-storage-0" (UID: "f94cf23a-424d-4bd6-8e19-b1f3dc7a0160") : configmap "swift-ring-files" not found Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.811179 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.818025 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.834553 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v8qj\" (UniqueName: \"kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-kube-api-access-8v8qj\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:30 crc kubenswrapper[4872]: I0126 09:22:30.836274 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.028081 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-nvsn5"] Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.029786 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.035614 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.040914 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-nvsn5"] Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.103878 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.104551 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.110451 4872 generic.go:334] "Generic (PLEG): container finished" podID="2b46a9b4-7d84-46a7-82f2-6093f2563f51" containerID="b3da9a795c7d3e2bc65b94657f324b46af54a0dac130519302eee884ddf18ded" exitCode=0 Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.110516 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" event={"ID":"2b46a9b4-7d84-46a7-82f2-6093f2563f51","Type":"ContainerDied","Data":"b3da9a795c7d3e2bc65b94657f324b46af54a0dac130519302eee884ddf18ded"} Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.110556 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" event={"ID":"2b46a9b4-7d84-46a7-82f2-6093f2563f51","Type":"ContainerStarted","Data":"8a35796980197b34a5c4bca0dd090c652c3902052ec3925930778738e1dfc28c"} Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.116629 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-etc-swift\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.116698 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-ring-data-devices\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.116732 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-scripts\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.116771 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pv58\" (UniqueName: \"kubernetes.io/projected/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-kube-api-access-2pv58\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.116823 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-dispersionconf\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.116871 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-combined-ca-bundle\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.116907 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-swiftconf\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.198160 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35971ebc-86cc-4a97-b3ba-ebe63c6f1200" path="/var/lib/kubelet/pods/35971ebc-86cc-4a97-b3ba-ebe63c6f1200/volumes" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.218139 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-etc-swift\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.218207 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-ring-data-devices\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.218232 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-scripts\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.218264 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pv58\" (UniqueName: \"kubernetes.io/projected/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-kube-api-access-2pv58\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.219365 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-etc-swift\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.219619 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-dispersionconf\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.219685 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-combined-ca-bundle\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.219732 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-swiftconf\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.219989 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-ring-data-devices\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.220367 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-scripts\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.223478 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-dispersionconf\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.224333 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-swiftconf\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.224337 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-combined-ca-bundle\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.240357 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pv58\" (UniqueName: \"kubernetes.io/projected/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-kube-api-access-2pv58\") pod \"swift-ring-rebalance-nvsn5\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.321416 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:31 crc kubenswrapper[4872]: E0126 09:22:31.321674 4872 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 26 09:22:31 crc kubenswrapper[4872]: E0126 09:22:31.321710 4872 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 26 09:22:31 crc kubenswrapper[4872]: E0126 09:22:31.321775 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift podName:f94cf23a-424d-4bd6-8e19-b1f3dc7a0160 nodeName:}" failed. No retries permitted until 2026-01-26 09:22:32.321755373 +0000 UTC m=+885.630595174 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift") pod "swift-storage-0" (UID: "f94cf23a-424d-4bd6-8e19-b1f3dc7a0160") : configmap "swift-ring-files" not found Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.489054 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:31 crc kubenswrapper[4872]: I0126 09:22:31.987359 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-nvsn5"] Jan 26 09:22:32 crc kubenswrapper[4872]: I0126 09:22:32.121998 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" event={"ID":"2b46a9b4-7d84-46a7-82f2-6093f2563f51","Type":"ContainerStarted","Data":"170b0b54295a8f77b4481faa61c311ebd1aef59712316d9bd22460b5e5708012"} Jan 26 09:22:32 crc kubenswrapper[4872]: I0126 09:22:32.123312 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:32 crc kubenswrapper[4872]: I0126 09:22:32.124979 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nvsn5" event={"ID":"3ffefe20-1e04-4a87-b2cf-5140aefbb87d","Type":"ContainerStarted","Data":"a5f26b5adc59bc27b8fa79fdb014a6a440f8c752655874eb073762d413c9237e"} Jan 26 09:22:32 crc kubenswrapper[4872]: I0126 09:22:32.173093 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" podStartSLOduration=3.173067132 podStartE2EDuration="3.173067132s" podCreationTimestamp="2026-01-26 09:22:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:22:32.156525315 +0000 UTC m=+885.465365126" watchObservedRunningTime="2026-01-26 09:22:32.173067132 +0000 UTC m=+885.481906933" Jan 26 09:22:32 crc kubenswrapper[4872]: I0126 09:22:32.373969 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:32 crc kubenswrapper[4872]: E0126 09:22:32.374212 4872 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 26 09:22:32 crc kubenswrapper[4872]: E0126 09:22:32.374470 4872 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 26 09:22:32 crc kubenswrapper[4872]: E0126 09:22:32.374531 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift podName:f94cf23a-424d-4bd6-8e19-b1f3dc7a0160 nodeName:}" failed. No retries permitted until 2026-01-26 09:22:34.374510885 +0000 UTC m=+887.683350686 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift") pod "swift-storage-0" (UID: "f94cf23a-424d-4bd6-8e19-b1f3dc7a0160") : configmap "swift-ring-files" not found Jan 26 09:22:34 crc kubenswrapper[4872]: I0126 09:22:34.424775 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:34 crc kubenswrapper[4872]: E0126 09:22:34.425265 4872 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 26 09:22:34 crc kubenswrapper[4872]: E0126 09:22:34.425343 4872 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 26 09:22:34 crc kubenswrapper[4872]: E0126 09:22:34.425449 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift podName:f94cf23a-424d-4bd6-8e19-b1f3dc7a0160 nodeName:}" failed. No retries permitted until 2026-01-26 09:22:38.425420604 +0000 UTC m=+891.734260405 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift") pod "swift-storage-0" (UID: "f94cf23a-424d-4bd6-8e19-b1f3dc7a0160") : configmap "swift-ring-files" not found Jan 26 09:22:36 crc kubenswrapper[4872]: I0126 09:22:36.099721 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Jan 26 09:22:36 crc kubenswrapper[4872]: I0126 09:22:36.100724 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Jan 26 09:22:36 crc kubenswrapper[4872]: I0126 09:22:36.184497 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nvsn5" event={"ID":"3ffefe20-1e04-4a87-b2cf-5140aefbb87d","Type":"ContainerStarted","Data":"00b66d2a3b87df557d2d995293ecd78226ba52338e3237259eda5e836c1466a1"} Jan 26 09:22:36 crc kubenswrapper[4872]: I0126 09:22:36.217414 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-nvsn5" podStartSLOduration=1.5008373659999998 podStartE2EDuration="5.217387971s" podCreationTimestamp="2026-01-26 09:22:31 +0000 UTC" firstStartedPulling="2026-01-26 09:22:31.998131156 +0000 UTC m=+885.306970957" lastFinishedPulling="2026-01-26 09:22:35.714681761 +0000 UTC m=+889.023521562" observedRunningTime="2026-01-26 09:22:36.217235897 +0000 UTC m=+889.526075708" watchObservedRunningTime="2026-01-26 09:22:36.217387971 +0000 UTC m=+889.526227772" Jan 26 09:22:36 crc kubenswrapper[4872]: I0126 09:22:36.473888 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Jan 26 09:22:36 crc kubenswrapper[4872]: I0126 09:22:36.561595 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.054524 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-76ba-account-create-update-4dklb"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.056872 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-76ba-account-create-update-4dklb" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.071565 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.074053 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-76ba-account-create-update-4dklb"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.138101 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-j82b4"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.139430 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j82b4" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.146547 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-j82b4"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.181121 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k87kq\" (UniqueName: \"kubernetes.io/projected/1717069a-a49e-44ae-bb02-ebc6265e44b3-kube-api-access-k87kq\") pod \"keystone-76ba-account-create-update-4dklb\" (UID: \"1717069a-a49e-44ae-bb02-ebc6265e44b3\") " pod="openstack/keystone-76ba-account-create-update-4dklb" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.181632 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1717069a-a49e-44ae-bb02-ebc6265e44b3-operator-scripts\") pod \"keystone-76ba-account-create-update-4dklb\" (UID: \"1717069a-a49e-44ae-bb02-ebc6265e44b3\") " pod="openstack/keystone-76ba-account-create-update-4dklb" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.241724 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-5x7dg"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.243775 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.247247 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.248730 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"204a2939-a385-4f7d-a08b-a008e6a14df0","Type":"ContainerStarted","Data":"8cc517b391a64d790edff26ef904dfda8f629a5cdf48b5b0dd0a7e787acc2e82"} Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.276872 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-5x7dg"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.284051 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k87kq\" (UniqueName: \"kubernetes.io/projected/1717069a-a49e-44ae-bb02-ebc6265e44b3-kube-api-access-k87kq\") pod \"keystone-76ba-account-create-update-4dklb\" (UID: \"1717069a-a49e-44ae-bb02-ebc6265e44b3\") " pod="openstack/keystone-76ba-account-create-update-4dklb" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.284188 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1717069a-a49e-44ae-bb02-ebc6265e44b3-operator-scripts\") pod \"keystone-76ba-account-create-update-4dklb\" (UID: \"1717069a-a49e-44ae-bb02-ebc6265e44b3\") " pod="openstack/keystone-76ba-account-create-update-4dklb" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.284314 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pswqc\" (UniqueName: \"kubernetes.io/projected/dfeb3d47-b023-4e41-9bdb-80241b5b0972-kube-api-access-pswqc\") pod \"keystone-db-create-j82b4\" (UID: \"dfeb3d47-b023-4e41-9bdb-80241b5b0972\") " pod="openstack/keystone-db-create-j82b4" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.284354 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfeb3d47-b023-4e41-9bdb-80241b5b0972-operator-scripts\") pod \"keystone-db-create-j82b4\" (UID: \"dfeb3d47-b023-4e41-9bdb-80241b5b0972\") " pod="openstack/keystone-db-create-j82b4" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.291551 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1717069a-a49e-44ae-bb02-ebc6265e44b3-operator-scripts\") pod \"keystone-76ba-account-create-update-4dklb\" (UID: \"1717069a-a49e-44ae-bb02-ebc6265e44b3\") " pod="openstack/keystone-76ba-account-create-update-4dklb" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.313736 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=20.945330029 podStartE2EDuration="52.313697929s" podCreationTimestamp="2026-01-26 09:21:45 +0000 UTC" firstStartedPulling="2026-01-26 09:22:05.478957364 +0000 UTC m=+858.787797155" lastFinishedPulling="2026-01-26 09:22:36.847325254 +0000 UTC m=+890.156165055" observedRunningTime="2026-01-26 09:22:37.307312979 +0000 UTC m=+890.616152780" watchObservedRunningTime="2026-01-26 09:22:37.313697929 +0000 UTC m=+890.622537730" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.346966 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k87kq\" (UniqueName: \"kubernetes.io/projected/1717069a-a49e-44ae-bb02-ebc6265e44b3-kube-api-access-k87kq\") pod \"keystone-76ba-account-create-update-4dklb\" (UID: \"1717069a-a49e-44ae-bb02-ebc6265e44b3\") " pod="openstack/keystone-76ba-account-create-update-4dklb" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.366597 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-x6t97"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.377022 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x6t97" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.387349 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-76ba-account-create-update-4dklb" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.388669 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2cdaf461-41f0-44d3-8fad-d852c276c343-ovn-rundir\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.388715 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cdaf461-41f0-44d3-8fad-d852c276c343-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.388746 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq5sr\" (UniqueName: \"kubernetes.io/projected/2cdaf461-41f0-44d3-8fad-d852c276c343-kube-api-access-tq5sr\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.388916 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2cdaf461-41f0-44d3-8fad-d852c276c343-ovs-rundir\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.388993 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pswqc\" (UniqueName: \"kubernetes.io/projected/dfeb3d47-b023-4e41-9bdb-80241b5b0972-kube-api-access-pswqc\") pod \"keystone-db-create-j82b4\" (UID: \"dfeb3d47-b023-4e41-9bdb-80241b5b0972\") " pod="openstack/keystone-db-create-j82b4" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.389048 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdaf461-41f0-44d3-8fad-d852c276c343-config\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.389127 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfeb3d47-b023-4e41-9bdb-80241b5b0972-operator-scripts\") pod \"keystone-db-create-j82b4\" (UID: \"dfeb3d47-b023-4e41-9bdb-80241b5b0972\") " pod="openstack/keystone-db-create-j82b4" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.389182 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cdaf461-41f0-44d3-8fad-d852c276c343-combined-ca-bundle\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.401987 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-x6t97"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.413770 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfeb3d47-b023-4e41-9bdb-80241b5b0972-operator-scripts\") pod \"keystone-db-create-j82b4\" (UID: \"dfeb3d47-b023-4e41-9bdb-80241b5b0972\") " pod="openstack/keystone-db-create-j82b4" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.439934 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pswqc\" (UniqueName: \"kubernetes.io/projected/dfeb3d47-b023-4e41-9bdb-80241b5b0972-kube-api-access-pswqc\") pod \"keystone-db-create-j82b4\" (UID: \"dfeb3d47-b023-4e41-9bdb-80241b5b0972\") " pod="openstack/keystone-db-create-j82b4" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.457929 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f159-account-create-update-xxdzr"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.460908 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j82b4" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.461642 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f159-account-create-update-xxdzr" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.467314 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.487179 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f159-account-create-update-xxdzr"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.495217 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdaf461-41f0-44d3-8fad-d852c276c343-config\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.495277 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0723de91-066f-474b-b7fb-f5f252a830bb-operator-scripts\") pod \"placement-f159-account-create-update-xxdzr\" (UID: \"0723de91-066f-474b-b7fb-f5f252a830bb\") " pod="openstack/placement-f159-account-create-update-xxdzr" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.495326 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cdaf461-41f0-44d3-8fad-d852c276c343-combined-ca-bundle\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.495384 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2cdaf461-41f0-44d3-8fad-d852c276c343-ovn-rundir\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.495403 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/940d731a-47aa-40db-826b-7c1c7b23ee24-operator-scripts\") pod \"placement-db-create-x6t97\" (UID: \"940d731a-47aa-40db-826b-7c1c7b23ee24\") " pod="openstack/placement-db-create-x6t97" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.495421 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cdaf461-41f0-44d3-8fad-d852c276c343-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.495440 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq5sr\" (UniqueName: \"kubernetes.io/projected/2cdaf461-41f0-44d3-8fad-d852c276c343-kube-api-access-tq5sr\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.495464 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jn5x\" (UniqueName: \"kubernetes.io/projected/940d731a-47aa-40db-826b-7c1c7b23ee24-kube-api-access-5jn5x\") pod \"placement-db-create-x6t97\" (UID: \"940d731a-47aa-40db-826b-7c1c7b23ee24\") " pod="openstack/placement-db-create-x6t97" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.495525 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2cdaf461-41f0-44d3-8fad-d852c276c343-ovs-rundir\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.495551 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn9rz\" (UniqueName: \"kubernetes.io/projected/0723de91-066f-474b-b7fb-f5f252a830bb-kube-api-access-qn9rz\") pod \"placement-f159-account-create-update-xxdzr\" (UID: \"0723de91-066f-474b-b7fb-f5f252a830bb\") " pod="openstack/placement-f159-account-create-update-xxdzr" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.496276 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cdaf461-41f0-44d3-8fad-d852c276c343-config\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.502761 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/2cdaf461-41f0-44d3-8fad-d852c276c343-ovn-rundir\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.503239 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/2cdaf461-41f0-44d3-8fad-d852c276c343-ovs-rundir\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.505895 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cdaf461-41f0-44d3-8fad-d852c276c343-combined-ca-bundle\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.511391 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cdaf461-41f0-44d3-8fad-d852c276c343-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.560225 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-rm64t"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.560589 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" podUID="2b46a9b4-7d84-46a7-82f2-6093f2563f51" containerName="dnsmasq-dns" containerID="cri-o://170b0b54295a8f77b4481faa61c311ebd1aef59712316d9bd22460b5e5708012" gracePeriod=10 Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.563111 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.584305 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq5sr\" (UniqueName: \"kubernetes.io/projected/2cdaf461-41f0-44d3-8fad-d852c276c343-kube-api-access-tq5sr\") pod \"ovn-controller-metrics-5x7dg\" (UID: \"2cdaf461-41f0-44d3-8fad-d852c276c343\") " pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.586921 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.587521 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.596594 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-5x7dg" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.597190 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/940d731a-47aa-40db-826b-7c1c7b23ee24-operator-scripts\") pod \"placement-db-create-x6t97\" (UID: \"940d731a-47aa-40db-826b-7c1c7b23ee24\") " pod="openstack/placement-db-create-x6t97" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.597246 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jn5x\" (UniqueName: \"kubernetes.io/projected/940d731a-47aa-40db-826b-7c1c7b23ee24-kube-api-access-5jn5x\") pod \"placement-db-create-x6t97\" (UID: \"940d731a-47aa-40db-826b-7c1c7b23ee24\") " pod="openstack/placement-db-create-x6t97" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.597324 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn9rz\" (UniqueName: \"kubernetes.io/projected/0723de91-066f-474b-b7fb-f5f252a830bb-kube-api-access-qn9rz\") pod \"placement-f159-account-create-update-xxdzr\" (UID: \"0723de91-066f-474b-b7fb-f5f252a830bb\") " pod="openstack/placement-f159-account-create-update-xxdzr" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.597380 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0723de91-066f-474b-b7fb-f5f252a830bb-operator-scripts\") pod \"placement-f159-account-create-update-xxdzr\" (UID: \"0723de91-066f-474b-b7fb-f5f252a830bb\") " pod="openstack/placement-f159-account-create-update-xxdzr" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.598181 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0723de91-066f-474b-b7fb-f5f252a830bb-operator-scripts\") pod \"placement-f159-account-create-update-xxdzr\" (UID: \"0723de91-066f-474b-b7fb-f5f252a830bb\") " pod="openstack/placement-f159-account-create-update-xxdzr" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.598867 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/940d731a-47aa-40db-826b-7c1c7b23ee24-operator-scripts\") pod \"placement-db-create-x6t97\" (UID: \"940d731a-47aa-40db-826b-7c1c7b23ee24\") " pod="openstack/placement-db-create-x6t97" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.610681 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-xtlb8"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.612237 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.638616 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.679321 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jn5x\" (UniqueName: \"kubernetes.io/projected/940d731a-47aa-40db-826b-7c1c7b23ee24-kube-api-access-5jn5x\") pod \"placement-db-create-x6t97\" (UID: \"940d731a-47aa-40db-826b-7c1c7b23ee24\") " pod="openstack/placement-db-create-x6t97" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.719123 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x6t97" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.725950 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck2nl\" (UniqueName: \"kubernetes.io/projected/0b762f55-df95-4e63-afdd-e41218cb6fae-kube-api-access-ck2nl\") pod \"dnsmasq-dns-74f6f696b9-xtlb8\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.726277 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-xtlb8\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.726351 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-xtlb8\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.726520 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-config\") pod \"dnsmasq-dns-74f6f696b9-xtlb8\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.737566 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-xtlb8"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.743333 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn9rz\" (UniqueName: \"kubernetes.io/projected/0723de91-066f-474b-b7fb-f5f252a830bb-kube-api-access-qn9rz\") pod \"placement-f159-account-create-update-xxdzr\" (UID: \"0723de91-066f-474b-b7fb-f5f252a830bb\") " pod="openstack/placement-f159-account-create-update-xxdzr" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.791435 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-h7qsd"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.797668 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-h7qsd" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.826548 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f159-account-create-update-xxdzr" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.831140 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-config\") pod \"dnsmasq-dns-74f6f696b9-xtlb8\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.836312 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-config\") pod \"dnsmasq-dns-74f6f696b9-xtlb8\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.863140 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck2nl\" (UniqueName: \"kubernetes.io/projected/0b762f55-df95-4e63-afdd-e41218cb6fae-kube-api-access-ck2nl\") pod \"dnsmasq-dns-74f6f696b9-xtlb8\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.863355 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-xtlb8\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.863417 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-xtlb8\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.864538 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-dns-svc\") pod \"dnsmasq-dns-74f6f696b9-xtlb8\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.869627 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-h7qsd"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.869927 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6f696b9-xtlb8\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.875268 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.903297 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-67f9-account-create-update-d2zpv"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.907898 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-67f9-account-create-update-d2zpv" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.932708 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-67f9-account-create-update-d2zpv"] Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.932843 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.950683 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-xtlb8"] Jan 26 09:22:37 crc kubenswrapper[4872]: E0126 09:22:37.951715 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-ck2nl], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" podUID="0b762f55-df95-4e63-afdd-e41218cb6fae" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.952429 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck2nl\" (UniqueName: \"kubernetes.io/projected/0b762f55-df95-4e63-afdd-e41218cb6fae-kube-api-access-ck2nl\") pod \"dnsmasq-dns-74f6f696b9-xtlb8\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.965476 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nr4t\" (UniqueName: \"kubernetes.io/projected/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a-kube-api-access-8nr4t\") pod \"glance-db-create-h7qsd\" (UID: \"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a\") " pod="openstack/glance-db-create-h7qsd" Jan 26 09:22:37 crc kubenswrapper[4872]: I0126 09:22:37.965559 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a-operator-scripts\") pod \"glance-db-create-h7qsd\" (UID: \"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a\") " pod="openstack/glance-db-create-h7qsd" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.003884 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-qdxbg"] Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.005691 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.014157 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.029607 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qdxbg"] Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.072000 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-config\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.072086 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.072117 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flm2k\" (UniqueName: \"kubernetes.io/projected/52cc682d-f3a9-47fc-9850-3019549c5f18-kube-api-access-flm2k\") pod \"glance-67f9-account-create-update-d2zpv\" (UID: \"52cc682d-f3a9-47fc-9850-3019549c5f18\") " pod="openstack/glance-67f9-account-create-update-d2zpv" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.072158 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-dns-svc\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.072180 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.072228 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t96pc\" (UniqueName: \"kubernetes.io/projected/9ba3969e-0cea-4a89-8092-b2cf63786ecf-kube-api-access-t96pc\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.072257 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nr4t\" (UniqueName: \"kubernetes.io/projected/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a-kube-api-access-8nr4t\") pod \"glance-db-create-h7qsd\" (UID: \"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a\") " pod="openstack/glance-db-create-h7qsd" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.072274 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52cc682d-f3a9-47fc-9850-3019549c5f18-operator-scripts\") pod \"glance-67f9-account-create-update-d2zpv\" (UID: \"52cc682d-f3a9-47fc-9850-3019549c5f18\") " pod="openstack/glance-67f9-account-create-update-d2zpv" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.072306 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a-operator-scripts\") pod \"glance-db-create-h7qsd\" (UID: \"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a\") " pod="openstack/glance-db-create-h7qsd" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.073463 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a-operator-scripts\") pod \"glance-db-create-h7qsd\" (UID: \"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a\") " pod="openstack/glance-db-create-h7qsd" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.133066 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nr4t\" (UniqueName: \"kubernetes.io/projected/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a-kube-api-access-8nr4t\") pod \"glance-db-create-h7qsd\" (UID: \"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a\") " pod="openstack/glance-db-create-h7qsd" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.177132 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-config\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.177231 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.177262 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flm2k\" (UniqueName: \"kubernetes.io/projected/52cc682d-f3a9-47fc-9850-3019549c5f18-kube-api-access-flm2k\") pod \"glance-67f9-account-create-update-d2zpv\" (UID: \"52cc682d-f3a9-47fc-9850-3019549c5f18\") " pod="openstack/glance-67f9-account-create-update-d2zpv" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.177306 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-dns-svc\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.177333 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.177378 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t96pc\" (UniqueName: \"kubernetes.io/projected/9ba3969e-0cea-4a89-8092-b2cf63786ecf-kube-api-access-t96pc\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.177424 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52cc682d-f3a9-47fc-9850-3019549c5f18-operator-scripts\") pod \"glance-67f9-account-create-update-d2zpv\" (UID: \"52cc682d-f3a9-47fc-9850-3019549c5f18\") " pod="openstack/glance-67f9-account-create-update-d2zpv" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.178589 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.179048 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.183954 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-dns-svc\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.187211 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52cc682d-f3a9-47fc-9850-3019549c5f18-operator-scripts\") pod \"glance-67f9-account-create-update-d2zpv\" (UID: \"52cc682d-f3a9-47fc-9850-3019549c5f18\") " pod="openstack/glance-67f9-account-create-update-d2zpv" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.190667 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-config\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.207542 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flm2k\" (UniqueName: \"kubernetes.io/projected/52cc682d-f3a9-47fc-9850-3019549c5f18-kube-api-access-flm2k\") pod \"glance-67f9-account-create-update-d2zpv\" (UID: \"52cc682d-f3a9-47fc-9850-3019549c5f18\") " pod="openstack/glance-67f9-account-create-update-d2zpv" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.212700 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t96pc\" (UniqueName: \"kubernetes.io/projected/9ba3969e-0cea-4a89-8092-b2cf63786ecf-kube-api-access-t96pc\") pod \"dnsmasq-dns-698758b865-qdxbg\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.267220 4872 generic.go:334] "Generic (PLEG): container finished" podID="2b46a9b4-7d84-46a7-82f2-6093f2563f51" containerID="170b0b54295a8f77b4481faa61c311ebd1aef59712316d9bd22460b5e5708012" exitCode=0 Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.267338 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" event={"ID":"2b46a9b4-7d84-46a7-82f2-6093f2563f51","Type":"ContainerDied","Data":"170b0b54295a8f77b4481faa61c311ebd1aef59712316d9bd22460b5e5708012"} Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.268072 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.286559 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.369826 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.398412 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-h7qsd" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.458329 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-67f9-account-create-update-d2zpv" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.480494 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.487653 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-ovsdbserver-nb\") pod \"0b762f55-df95-4e63-afdd-e41218cb6fae\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.487990 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-dns-svc\") pod \"0b762f55-df95-4e63-afdd-e41218cb6fae\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.488356 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-config\") pod \"0b762f55-df95-4e63-afdd-e41218cb6fae\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.488943 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck2nl\" (UniqueName: \"kubernetes.io/projected/0b762f55-df95-4e63-afdd-e41218cb6fae-kube-api-access-ck2nl\") pod \"0b762f55-df95-4e63-afdd-e41218cb6fae\" (UID: \"0b762f55-df95-4e63-afdd-e41218cb6fae\") " Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.489283 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-config" (OuterVolumeSpecName: "config") pod "0b762f55-df95-4e63-afdd-e41218cb6fae" (UID: "0b762f55-df95-4e63-afdd-e41218cb6fae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.489409 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0b762f55-df95-4e63-afdd-e41218cb6fae" (UID: "0b762f55-df95-4e63-afdd-e41218cb6fae"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.489659 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.489993 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.490014 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:38 crc kubenswrapper[4872]: E0126 09:22:38.490227 4872 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Jan 26 09:22:38 crc kubenswrapper[4872]: E0126 09:22:38.490297 4872 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Jan 26 09:22:38 crc kubenswrapper[4872]: E0126 09:22:38.490402 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift podName:f94cf23a-424d-4bd6-8e19-b1f3dc7a0160 nodeName:}" failed. No retries permitted until 2026-01-26 09:22:46.490378072 +0000 UTC m=+899.799217873 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift") pod "swift-storage-0" (UID: "f94cf23a-424d-4bd6-8e19-b1f3dc7a0160") : configmap "swift-ring-files" not found Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.507246 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0b762f55-df95-4e63-afdd-e41218cb6fae" (UID: "0b762f55-df95-4e63-afdd-e41218cb6fae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.562258 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b762f55-df95-4e63-afdd-e41218cb6fae-kube-api-access-ck2nl" (OuterVolumeSpecName: "kube-api-access-ck2nl") pod "0b762f55-df95-4e63-afdd-e41218cb6fae" (UID: "0b762f55-df95-4e63-afdd-e41218cb6fae"). InnerVolumeSpecName "kube-api-access-ck2nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.592717 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b762f55-df95-4e63-afdd-e41218cb6fae-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.593580 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck2nl\" (UniqueName: \"kubernetes.io/projected/0b762f55-df95-4e63-afdd-e41218cb6fae-kube-api-access-ck2nl\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.678850 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-j82b4"] Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.682484 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.699596 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-76ba-account-create-update-4dklb"] Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.798690 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b46a9b4-7d84-46a7-82f2-6093f2563f51-dns-svc\") pod \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\" (UID: \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\") " Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.798862 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hskl\" (UniqueName: \"kubernetes.io/projected/2b46a9b4-7d84-46a7-82f2-6093f2563f51-kube-api-access-6hskl\") pod \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\" (UID: \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\") " Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.798949 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b46a9b4-7d84-46a7-82f2-6093f2563f51-config\") pod \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\" (UID: \"2b46a9b4-7d84-46a7-82f2-6093f2563f51\") " Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.826253 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b46a9b4-7d84-46a7-82f2-6093f2563f51-kube-api-access-6hskl" (OuterVolumeSpecName: "kube-api-access-6hskl") pod "2b46a9b4-7d84-46a7-82f2-6093f2563f51" (UID: "2b46a9b4-7d84-46a7-82f2-6093f2563f51"). InnerVolumeSpecName "kube-api-access-6hskl". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.895626 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b46a9b4-7d84-46a7-82f2-6093f2563f51-config" (OuterVolumeSpecName: "config") pod "2b46a9b4-7d84-46a7-82f2-6093f2563f51" (UID: "2b46a9b4-7d84-46a7-82f2-6093f2563f51"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.903415 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hskl\" (UniqueName: \"kubernetes.io/projected/2b46a9b4-7d84-46a7-82f2-6093f2563f51-kube-api-access-6hskl\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.903476 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2b46a9b4-7d84-46a7-82f2-6093f2563f51-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:38 crc kubenswrapper[4872]: I0126 09:22:38.966077 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b46a9b4-7d84-46a7-82f2-6093f2563f51-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2b46a9b4-7d84-46a7-82f2-6093f2563f51" (UID: "2b46a9b4-7d84-46a7-82f2-6093f2563f51"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.012514 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2b46a9b4-7d84-46a7-82f2-6093f2563f51-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.281412 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-76ba-account-create-update-4dklb" event={"ID":"1717069a-a49e-44ae-bb02-ebc6265e44b3","Type":"ContainerStarted","Data":"93cd57027a2c230439a971b6139de2451a4cf2f1ebf6f58303b5f661572649b7"} Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.294904 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" event={"ID":"2b46a9b4-7d84-46a7-82f2-6093f2563f51","Type":"ContainerDied","Data":"8a35796980197b34a5c4bca0dd090c652c3902052ec3925930778738e1dfc28c"} Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.295004 4872 scope.go:117] "RemoveContainer" containerID="170b0b54295a8f77b4481faa61c311ebd1aef59712316d9bd22460b5e5708012" Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.295194 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-rm64t" Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.308428 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j82b4" event={"ID":"dfeb3d47-b023-4e41-9bdb-80241b5b0972","Type":"ContainerStarted","Data":"54cc1719acb7c4737547f96ea3dbe07fba986059fba1d6de081740a6adef6f3f"} Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.308564 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6f696b9-xtlb8" Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.335993 4872 scope.go:117] "RemoveContainer" containerID="b3da9a795c7d3e2bc65b94657f324b46af54a0dac130519302eee884ddf18ded" Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.346066 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-5x7dg"] Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.380077 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f159-account-create-update-xxdzr"] Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.419787 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-x6t97"] Jan 26 09:22:39 crc kubenswrapper[4872]: W0126 09:22:39.432975 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0723de91_066f_474b_b7fb_f5f252a830bb.slice/crio-195ea6e4f398ccecb80d5c2d6a657f62b60fd9f90349220ca1d044f0ff0d1545 WatchSource:0}: Error finding container 195ea6e4f398ccecb80d5c2d6a657f62b60fd9f90349220ca1d044f0ff0d1545: Status 404 returned error can't find the container with id 195ea6e4f398ccecb80d5c2d6a657f62b60fd9f90349220ca1d044f0ff0d1545 Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.441128 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-xtlb8"] Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.451430 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6f696b9-xtlb8"] Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.466704 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-rm64t"] Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.470923 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-rm64t"] Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.700920 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-h7qsd"] Jan 26 09:22:39 crc kubenswrapper[4872]: W0126 09:22:39.714474 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod392ab1c3_7cbf_44ac_8b8e_39b64e502d8a.slice/crio-8dc17be4da60801f951f64255041ede57ab380ab4820c19ad3397a6d64d735d0 WatchSource:0}: Error finding container 8dc17be4da60801f951f64255041ede57ab380ab4820c19ad3397a6d64d735d0: Status 404 returned error can't find the container with id 8dc17be4da60801f951f64255041ede57ab380ab4820c19ad3397a6d64d735d0 Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.725383 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qdxbg"] Jan 26 09:22:39 crc kubenswrapper[4872]: W0126 09:22:39.746687 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ba3969e_0cea_4a89_8092_b2cf63786ecf.slice/crio-b6ddbf56dbf9c1716878ae51b75d3b11b9f41bbb9da617ef936fa09bce6536c2 WatchSource:0}: Error finding container b6ddbf56dbf9c1716878ae51b75d3b11b9f41bbb9da617ef936fa09bce6536c2: Status 404 returned error can't find the container with id b6ddbf56dbf9c1716878ae51b75d3b11b9f41bbb9da617ef936fa09bce6536c2 Jan 26 09:22:39 crc kubenswrapper[4872]: I0126 09:22:39.771489 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-67f9-account-create-update-d2zpv"] Jan 26 09:22:39 crc kubenswrapper[4872]: W0126 09:22:39.773470 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52cc682d_f3a9_47fc_9850_3019549c5f18.slice/crio-1653cff969c4ef4ef349c06c3790b83df82ede6849a531615c291414ae177699 WatchSource:0}: Error finding container 1653cff969c4ef4ef349c06c3790b83df82ede6849a531615c291414ae177699: Status 404 returned error can't find the container with id 1653cff969c4ef4ef349c06c3790b83df82ede6849a531615c291414ae177699 Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.317189 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qdxbg" event={"ID":"9ba3969e-0cea-4a89-8092-b2cf63786ecf","Type":"ContainerStarted","Data":"b6ddbf56dbf9c1716878ae51b75d3b11b9f41bbb9da617ef936fa09bce6536c2"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.319291 4872 generic.go:334] "Generic (PLEG): container finished" podID="52cc682d-f3a9-47fc-9850-3019549c5f18" containerID="e449be6901dedad52713a04f03ce664d0afa53c4eb9ea0b060eb6574f4fd2223" exitCode=0 Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.319357 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-67f9-account-create-update-d2zpv" event={"ID":"52cc682d-f3a9-47fc-9850-3019549c5f18","Type":"ContainerDied","Data":"e449be6901dedad52713a04f03ce664d0afa53c4eb9ea0b060eb6574f4fd2223"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.319381 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-67f9-account-create-update-d2zpv" event={"ID":"52cc682d-f3a9-47fc-9850-3019549c5f18","Type":"ContainerStarted","Data":"1653cff969c4ef4ef349c06c3790b83df82ede6849a531615c291414ae177699"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.322422 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"bf0544f2-34dc-45fc-986e-cb2554c04566","Type":"ContainerStarted","Data":"6947049a8463a4a587473309cdf1d5c456a99ace97762e03c9d3c6f9f07f7d82"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.326279 4872 generic.go:334] "Generic (PLEG): container finished" podID="1717069a-a49e-44ae-bb02-ebc6265e44b3" containerID="b6df7e8486927d214ef6662f821a83abaa172de17822d798e0e0c1bf0ef414f2" exitCode=0 Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.326354 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-76ba-account-create-update-4dklb" event={"ID":"1717069a-a49e-44ae-bb02-ebc6265e44b3","Type":"ContainerDied","Data":"b6df7e8486927d214ef6662f821a83abaa172de17822d798e0e0c1bf0ef414f2"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.328928 4872 generic.go:334] "Generic (PLEG): container finished" podID="392ab1c3-7cbf-44ac-8b8e-39b64e502d8a" containerID="8efba7583a85ca87b0a3874892d9272bf4c2553acbc8b2acc196569532f77bc7" exitCode=0 Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.329001 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-h7qsd" event={"ID":"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a","Type":"ContainerDied","Data":"8efba7583a85ca87b0a3874892d9272bf4c2553acbc8b2acc196569532f77bc7"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.329027 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-h7qsd" event={"ID":"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a","Type":"ContainerStarted","Data":"8dc17be4da60801f951f64255041ede57ab380ab4820c19ad3397a6d64d735d0"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.332174 4872 generic.go:334] "Generic (PLEG): container finished" podID="0723de91-066f-474b-b7fb-f5f252a830bb" containerID="fb54217da7329679731ceace5fdfe65aa5e1c5d9a7f57e23f7f53392e3580e8c" exitCode=0 Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.332353 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f159-account-create-update-xxdzr" event={"ID":"0723de91-066f-474b-b7fb-f5f252a830bb","Type":"ContainerDied","Data":"fb54217da7329679731ceace5fdfe65aa5e1c5d9a7f57e23f7f53392e3580e8c"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.332392 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f159-account-create-update-xxdzr" event={"ID":"0723de91-066f-474b-b7fb-f5f252a830bb","Type":"ContainerStarted","Data":"195ea6e4f398ccecb80d5c2d6a657f62b60fd9f90349220ca1d044f0ff0d1545"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.336833 4872 generic.go:334] "Generic (PLEG): container finished" podID="dfeb3d47-b023-4e41-9bdb-80241b5b0972" containerID="bf89441413031f9b78b00cf7b749940f7bccf809ad4cb47e9fb0a74c79b7b125" exitCode=0 Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.336909 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j82b4" event={"ID":"dfeb3d47-b023-4e41-9bdb-80241b5b0972","Type":"ContainerDied","Data":"bf89441413031f9b78b00cf7b749940f7bccf809ad4cb47e9fb0a74c79b7b125"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.338644 4872 generic.go:334] "Generic (PLEG): container finished" podID="940d731a-47aa-40db-826b-7c1c7b23ee24" containerID="b6c093bbea84de36a73de76ec30d497fa3c0bd401ed4f88003495408816cdeb4" exitCode=0 Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.338734 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-x6t97" event={"ID":"940d731a-47aa-40db-826b-7c1c7b23ee24","Type":"ContainerDied","Data":"b6c093bbea84de36a73de76ec30d497fa3c0bd401ed4f88003495408816cdeb4"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.338770 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-x6t97" event={"ID":"940d731a-47aa-40db-826b-7c1c7b23ee24","Type":"ContainerStarted","Data":"57c7022e80e6f682b86c70b60391d77e5f3f03a1d986493a1cda9d8a8b0ae11f"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.343496 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-5x7dg" event={"ID":"2cdaf461-41f0-44d3-8fad-d852c276c343","Type":"ContainerStarted","Data":"a91cc7de3c599b12d1c185c0f1178ac2a2818640188c1550a20b136d6c6fe2cd"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.343545 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-5x7dg" event={"ID":"2cdaf461-41f0-44d3-8fad-d852c276c343","Type":"ContainerStarted","Data":"fd3b0622bc9bcbf75c03dfe7b908689ef4ba9dab47f768358c0dbaf90d2cfd67"} Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.376827 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=54.29675139 podStartE2EDuration="58.376776078s" podCreationTimestamp="2026-01-26 09:21:42 +0000 UTC" firstStartedPulling="2026-01-26 09:22:04.550251806 +0000 UTC m=+857.859091607" lastFinishedPulling="2026-01-26 09:22:08.630276494 +0000 UTC m=+861.939116295" observedRunningTime="2026-01-26 09:22:40.373365792 +0000 UTC m=+893.682205593" watchObservedRunningTime="2026-01-26 09:22:40.376776078 +0000 UTC m=+893.685615889" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.527244 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-5x7dg" podStartSLOduration=3.527221397 podStartE2EDuration="3.527221397s" podCreationTimestamp="2026-01-26 09:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:22:40.523634006 +0000 UTC m=+893.832473817" watchObservedRunningTime="2026-01-26 09:22:40.527221397 +0000 UTC m=+893.836061188" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.629364 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Jan 26 09:22:40 crc kubenswrapper[4872]: E0126 09:22:40.629844 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b46a9b4-7d84-46a7-82f2-6093f2563f51" containerName="dnsmasq-dns" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.629859 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b46a9b4-7d84-46a7-82f2-6093f2563f51" containerName="dnsmasq-dns" Jan 26 09:22:40 crc kubenswrapper[4872]: E0126 09:22:40.629869 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b46a9b4-7d84-46a7-82f2-6093f2563f51" containerName="init" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.629875 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b46a9b4-7d84-46a7-82f2-6093f2563f51" containerName="init" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.630055 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b46a9b4-7d84-46a7-82f2-6093f2563f51" containerName="dnsmasq-dns" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.631301 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.636965 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.637713 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.637966 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-fjt5r" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.638846 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.653167 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.770781 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc8642bd-5341-4e95-be61-deb51d754edb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.771356 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc8642bd-5341-4e95-be61-deb51d754edb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.771402 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8642bd-5341-4e95-be61-deb51d754edb-config\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.771421 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc8642bd-5341-4e95-be61-deb51d754edb-scripts\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.771499 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8642bd-5341-4e95-be61-deb51d754edb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.771618 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fc8642bd-5341-4e95-be61-deb51d754edb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.771677 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28x8x\" (UniqueName: \"kubernetes.io/projected/fc8642bd-5341-4e95-be61-deb51d754edb-kube-api-access-28x8x\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.873650 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8642bd-5341-4e95-be61-deb51d754edb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.873795 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fc8642bd-5341-4e95-be61-deb51d754edb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.873850 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28x8x\" (UniqueName: \"kubernetes.io/projected/fc8642bd-5341-4e95-be61-deb51d754edb-kube-api-access-28x8x\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.873960 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc8642bd-5341-4e95-be61-deb51d754edb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.875339 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc8642bd-5341-4e95-be61-deb51d754edb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.875394 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8642bd-5341-4e95-be61-deb51d754edb-config\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.875422 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc8642bd-5341-4e95-be61-deb51d754edb-scripts\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.874512 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fc8642bd-5341-4e95-be61-deb51d754edb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.876642 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8642bd-5341-4e95-be61-deb51d754edb-config\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.876739 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc8642bd-5341-4e95-be61-deb51d754edb-scripts\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.882145 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc8642bd-5341-4e95-be61-deb51d754edb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.882209 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc8642bd-5341-4e95-be61-deb51d754edb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.894892 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28x8x\" (UniqueName: \"kubernetes.io/projected/fc8642bd-5341-4e95-be61-deb51d754edb-kube-api-access-28x8x\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:40 crc kubenswrapper[4872]: I0126 09:22:40.897293 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc8642bd-5341-4e95-be61-deb51d754edb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"fc8642bd-5341-4e95-be61-deb51d754edb\") " pod="openstack/ovn-northd-0" Jan 26 09:22:41 crc kubenswrapper[4872]: I0126 09:22:41.006240 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Jan 26 09:22:41 crc kubenswrapper[4872]: I0126 09:22:41.203547 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b762f55-df95-4e63-afdd-e41218cb6fae" path="/var/lib/kubelet/pods/0b762f55-df95-4e63-afdd-e41218cb6fae/volumes" Jan 26 09:22:41 crc kubenswrapper[4872]: I0126 09:22:41.204255 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b46a9b4-7d84-46a7-82f2-6093f2563f51" path="/var/lib/kubelet/pods/2b46a9b4-7d84-46a7-82f2-6093f2563f51/volumes" Jan 26 09:22:41 crc kubenswrapper[4872]: I0126 09:22:41.358689 4872 generic.go:334] "Generic (PLEG): container finished" podID="9ba3969e-0cea-4a89-8092-b2cf63786ecf" containerID="708bdc6a21183df94d75a797e55f60f737d9cd47ec1c48e5347d0d550f485a62" exitCode=0 Jan 26 09:22:41 crc kubenswrapper[4872]: I0126 09:22:41.360307 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qdxbg" event={"ID":"9ba3969e-0cea-4a89-8092-b2cf63786ecf","Type":"ContainerDied","Data":"708bdc6a21183df94d75a797e55f60f737d9cd47ec1c48e5347d0d550f485a62"} Jan 26 09:22:41 crc kubenswrapper[4872]: I0126 09:22:41.490817 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.098186 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x6t97" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.206093 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/940d731a-47aa-40db-826b-7c1c7b23ee24-operator-scripts\") pod \"940d731a-47aa-40db-826b-7c1c7b23ee24\" (UID: \"940d731a-47aa-40db-826b-7c1c7b23ee24\") " Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.206326 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jn5x\" (UniqueName: \"kubernetes.io/projected/940d731a-47aa-40db-826b-7c1c7b23ee24-kube-api-access-5jn5x\") pod \"940d731a-47aa-40db-826b-7c1c7b23ee24\" (UID: \"940d731a-47aa-40db-826b-7c1c7b23ee24\") " Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.207261 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/940d731a-47aa-40db-826b-7c1c7b23ee24-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "940d731a-47aa-40db-826b-7c1c7b23ee24" (UID: "940d731a-47aa-40db-826b-7c1c7b23ee24"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.229210 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/940d731a-47aa-40db-826b-7c1c7b23ee24-kube-api-access-5jn5x" (OuterVolumeSpecName: "kube-api-access-5jn5x") pod "940d731a-47aa-40db-826b-7c1c7b23ee24" (UID: "940d731a-47aa-40db-826b-7c1c7b23ee24"). InnerVolumeSpecName "kube-api-access-5jn5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.308968 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/940d731a-47aa-40db-826b-7c1c7b23ee24-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.309030 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jn5x\" (UniqueName: \"kubernetes.io/projected/940d731a-47aa-40db-826b-7c1c7b23ee24-kube-api-access-5jn5x\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.381352 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qdxbg" event={"ID":"9ba3969e-0cea-4a89-8092-b2cf63786ecf","Type":"ContainerStarted","Data":"14a8d88b6802e3b097ff366b7cd26b9edda8faaa71177069aef9dc53ca1ce533"} Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.381605 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.395773 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-67f9-account-create-update-d2zpv" event={"ID":"52cc682d-f3a9-47fc-9850-3019549c5f18","Type":"ContainerDied","Data":"1653cff969c4ef4ef349c06c3790b83df82ede6849a531615c291414ae177699"} Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.395837 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1653cff969c4ef4ef349c06c3790b83df82ede6849a531615c291414ae177699" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.397265 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-x6t97" event={"ID":"940d731a-47aa-40db-826b-7c1c7b23ee24","Type":"ContainerDied","Data":"57c7022e80e6f682b86c70b60391d77e5f3f03a1d986493a1cda9d8a8b0ae11f"} Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.397323 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57c7022e80e6f682b86c70b60391d77e5f3f03a1d986493a1cda9d8a8b0ae11f" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.397876 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-x6t97" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.401172 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fc8642bd-5341-4e95-be61-deb51d754edb","Type":"ContainerStarted","Data":"09edbacccdbfa464c5871a74f8340674d9cdc25def7fa5af769005b76f28f5d2"} Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.401535 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-67f9-account-create-update-d2zpv" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.416646 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-qdxbg" podStartSLOduration=5.416623768 podStartE2EDuration="5.416623768s" podCreationTimestamp="2026-01-26 09:22:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:22:42.409826527 +0000 UTC m=+895.718666328" watchObservedRunningTime="2026-01-26 09:22:42.416623768 +0000 UTC m=+895.725463569" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.417374 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-76ba-account-create-update-4dklb" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.423502 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j82b4" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.449020 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f159-account-create-update-xxdzr" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.468965 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-h7qsd" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.517785 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1717069a-a49e-44ae-bb02-ebc6265e44b3-operator-scripts\") pod \"1717069a-a49e-44ae-bb02-ebc6265e44b3\" (UID: \"1717069a-a49e-44ae-bb02-ebc6265e44b3\") " Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.517903 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qn9rz\" (UniqueName: \"kubernetes.io/projected/0723de91-066f-474b-b7fb-f5f252a830bb-kube-api-access-qn9rz\") pod \"0723de91-066f-474b-b7fb-f5f252a830bb\" (UID: \"0723de91-066f-474b-b7fb-f5f252a830bb\") " Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.517946 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nr4t\" (UniqueName: \"kubernetes.io/projected/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a-kube-api-access-8nr4t\") pod \"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a\" (UID: \"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a\") " Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.517972 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flm2k\" (UniqueName: \"kubernetes.io/projected/52cc682d-f3a9-47fc-9850-3019549c5f18-kube-api-access-flm2k\") pod \"52cc682d-f3a9-47fc-9850-3019549c5f18\" (UID: \"52cc682d-f3a9-47fc-9850-3019549c5f18\") " Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.518087 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52cc682d-f3a9-47fc-9850-3019549c5f18-operator-scripts\") pod \"52cc682d-f3a9-47fc-9850-3019549c5f18\" (UID: \"52cc682d-f3a9-47fc-9850-3019549c5f18\") " Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.518165 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0723de91-066f-474b-b7fb-f5f252a830bb-operator-scripts\") pod \"0723de91-066f-474b-b7fb-f5f252a830bb\" (UID: \"0723de91-066f-474b-b7fb-f5f252a830bb\") " Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.518227 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a-operator-scripts\") pod \"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a\" (UID: \"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a\") " Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.518322 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k87kq\" (UniqueName: \"kubernetes.io/projected/1717069a-a49e-44ae-bb02-ebc6265e44b3-kube-api-access-k87kq\") pod \"1717069a-a49e-44ae-bb02-ebc6265e44b3\" (UID: \"1717069a-a49e-44ae-bb02-ebc6265e44b3\") " Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.518440 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pswqc\" (UniqueName: \"kubernetes.io/projected/dfeb3d47-b023-4e41-9bdb-80241b5b0972-kube-api-access-pswqc\") pod \"dfeb3d47-b023-4e41-9bdb-80241b5b0972\" (UID: \"dfeb3d47-b023-4e41-9bdb-80241b5b0972\") " Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.518477 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfeb3d47-b023-4e41-9bdb-80241b5b0972-operator-scripts\") pod \"dfeb3d47-b023-4e41-9bdb-80241b5b0972\" (UID: \"dfeb3d47-b023-4e41-9bdb-80241b5b0972\") " Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.519144 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0723de91-066f-474b-b7fb-f5f252a830bb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0723de91-066f-474b-b7fb-f5f252a830bb" (UID: "0723de91-066f-474b-b7fb-f5f252a830bb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.519715 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52cc682d-f3a9-47fc-9850-3019549c5f18-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "52cc682d-f3a9-47fc-9850-3019549c5f18" (UID: "52cc682d-f3a9-47fc-9850-3019549c5f18"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.519734 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "392ab1c3-7cbf-44ac-8b8e-39b64e502d8a" (UID: "392ab1c3-7cbf-44ac-8b8e-39b64e502d8a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.519739 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfeb3d47-b023-4e41-9bdb-80241b5b0972-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dfeb3d47-b023-4e41-9bdb-80241b5b0972" (UID: "dfeb3d47-b023-4e41-9bdb-80241b5b0972"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.521637 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1717069a-a49e-44ae-bb02-ebc6265e44b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1717069a-a49e-44ae-bb02-ebc6265e44b3" (UID: "1717069a-a49e-44ae-bb02-ebc6265e44b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.527865 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52cc682d-f3a9-47fc-9850-3019549c5f18-kube-api-access-flm2k" (OuterVolumeSpecName: "kube-api-access-flm2k") pod "52cc682d-f3a9-47fc-9850-3019549c5f18" (UID: "52cc682d-f3a9-47fc-9850-3019549c5f18"). InnerVolumeSpecName "kube-api-access-flm2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.529413 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfeb3d47-b023-4e41-9bdb-80241b5b0972-kube-api-access-pswqc" (OuterVolumeSpecName: "kube-api-access-pswqc") pod "dfeb3d47-b023-4e41-9bdb-80241b5b0972" (UID: "dfeb3d47-b023-4e41-9bdb-80241b5b0972"). InnerVolumeSpecName "kube-api-access-pswqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.534235 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1717069a-a49e-44ae-bb02-ebc6265e44b3-kube-api-access-k87kq" (OuterVolumeSpecName: "kube-api-access-k87kq") pod "1717069a-a49e-44ae-bb02-ebc6265e44b3" (UID: "1717069a-a49e-44ae-bb02-ebc6265e44b3"). InnerVolumeSpecName "kube-api-access-k87kq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.534308 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0723de91-066f-474b-b7fb-f5f252a830bb-kube-api-access-qn9rz" (OuterVolumeSpecName: "kube-api-access-qn9rz") pod "0723de91-066f-474b-b7fb-f5f252a830bb" (UID: "0723de91-066f-474b-b7fb-f5f252a830bb"). InnerVolumeSpecName "kube-api-access-qn9rz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.534471 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a-kube-api-access-8nr4t" (OuterVolumeSpecName: "kube-api-access-8nr4t") pod "392ab1c3-7cbf-44ac-8b8e-39b64e502d8a" (UID: "392ab1c3-7cbf-44ac-8b8e-39b64e502d8a"). InnerVolumeSpecName "kube-api-access-8nr4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.620973 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/52cc682d-f3a9-47fc-9850-3019549c5f18-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.621022 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0723de91-066f-474b-b7fb-f5f252a830bb-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.621035 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.621048 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k87kq\" (UniqueName: \"kubernetes.io/projected/1717069a-a49e-44ae-bb02-ebc6265e44b3-kube-api-access-k87kq\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.621060 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pswqc\" (UniqueName: \"kubernetes.io/projected/dfeb3d47-b023-4e41-9bdb-80241b5b0972-kube-api-access-pswqc\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.621077 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dfeb3d47-b023-4e41-9bdb-80241b5b0972-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.621087 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1717069a-a49e-44ae-bb02-ebc6265e44b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.621097 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qn9rz\" (UniqueName: \"kubernetes.io/projected/0723de91-066f-474b-b7fb-f5f252a830bb-kube-api-access-qn9rz\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.621113 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flm2k\" (UniqueName: \"kubernetes.io/projected/52cc682d-f3a9-47fc-9850-3019549c5f18-kube-api-access-flm2k\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:42 crc kubenswrapper[4872]: I0126 09:22:42.621122 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nr4t\" (UniqueName: \"kubernetes.io/projected/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a-kube-api-access-8nr4t\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.169710 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-gmp2c" podUID="cb72fdef-372c-41a1-bf57-6c71162ab194" containerName="ovn-controller" probeResult="failure" output=< Jan 26 09:22:43 crc kubenswrapper[4872]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Jan 26 09:22:43 crc kubenswrapper[4872]: > Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.275111 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.276277 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mgn5r" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.431513 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fc8642bd-5341-4e95-be61-deb51d754edb","Type":"ContainerStarted","Data":"aca4c3e6d657df10ac2d7d1f8df9277e2e2e50a072395a7cc1d7d109d6bfd76f"} Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.433334 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-h7qsd" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.433455 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-h7qsd" event={"ID":"392ab1c3-7cbf-44ac-8b8e-39b64e502d8a","Type":"ContainerDied","Data":"8dc17be4da60801f951f64255041ede57ab380ab4820c19ad3397a6d64d735d0"} Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.433554 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dc17be4da60801f951f64255041ede57ab380ab4820c19ad3397a6d64d735d0" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.443433 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f159-account-create-update-xxdzr" event={"ID":"0723de91-066f-474b-b7fb-f5f252a830bb","Type":"ContainerDied","Data":"195ea6e4f398ccecb80d5c2d6a657f62b60fd9f90349220ca1d044f0ff0d1545"} Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.443474 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="195ea6e4f398ccecb80d5c2d6a657f62b60fd9f90349220ca1d044f0ff0d1545" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.443486 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f159-account-create-update-xxdzr" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.449176 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j82b4" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.449489 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j82b4" event={"ID":"dfeb3d47-b023-4e41-9bdb-80241b5b0972","Type":"ContainerDied","Data":"54cc1719acb7c4737547f96ea3dbe07fba986059fba1d6de081740a6adef6f3f"} Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.449537 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54cc1719acb7c4737547f96ea3dbe07fba986059fba1d6de081740a6adef6f3f" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.457791 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-76ba-account-create-update-4dklb" event={"ID":"1717069a-a49e-44ae-bb02-ebc6265e44b3","Type":"ContainerDied","Data":"93cd57027a2c230439a971b6139de2451a4cf2f1ebf6f58303b5f661572649b7"} Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.457867 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-67f9-account-create-update-d2zpv" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.457875 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93cd57027a2c230439a971b6139de2451a4cf2f1ebf6f58303b5f661572649b7" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.457957 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-76ba-account-create-update-4dklb" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.558688 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gmp2c-config-vsj5m"] Jan 26 09:22:43 crc kubenswrapper[4872]: E0126 09:22:43.559171 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="392ab1c3-7cbf-44ac-8b8e-39b64e502d8a" containerName="mariadb-database-create" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.559198 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="392ab1c3-7cbf-44ac-8b8e-39b64e502d8a" containerName="mariadb-database-create" Jan 26 09:22:43 crc kubenswrapper[4872]: E0126 09:22:43.559224 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="940d731a-47aa-40db-826b-7c1c7b23ee24" containerName="mariadb-database-create" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.559233 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="940d731a-47aa-40db-826b-7c1c7b23ee24" containerName="mariadb-database-create" Jan 26 09:22:43 crc kubenswrapper[4872]: E0126 09:22:43.559251 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1717069a-a49e-44ae-bb02-ebc6265e44b3" containerName="mariadb-account-create-update" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.559261 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1717069a-a49e-44ae-bb02-ebc6265e44b3" containerName="mariadb-account-create-update" Jan 26 09:22:43 crc kubenswrapper[4872]: E0126 09:22:43.559273 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfeb3d47-b023-4e41-9bdb-80241b5b0972" containerName="mariadb-database-create" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.559283 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfeb3d47-b023-4e41-9bdb-80241b5b0972" containerName="mariadb-database-create" Jan 26 09:22:43 crc kubenswrapper[4872]: E0126 09:22:43.559300 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0723de91-066f-474b-b7fb-f5f252a830bb" containerName="mariadb-account-create-update" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.559309 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0723de91-066f-474b-b7fb-f5f252a830bb" containerName="mariadb-account-create-update" Jan 26 09:22:43 crc kubenswrapper[4872]: E0126 09:22:43.559320 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52cc682d-f3a9-47fc-9850-3019549c5f18" containerName="mariadb-account-create-update" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.559328 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="52cc682d-f3a9-47fc-9850-3019549c5f18" containerName="mariadb-account-create-update" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.559532 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="1717069a-a49e-44ae-bb02-ebc6265e44b3" containerName="mariadb-account-create-update" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.559553 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="52cc682d-f3a9-47fc-9850-3019549c5f18" containerName="mariadb-account-create-update" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.559567 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfeb3d47-b023-4e41-9bdb-80241b5b0972" containerName="mariadb-database-create" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.559577 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="392ab1c3-7cbf-44ac-8b8e-39b64e502d8a" containerName="mariadb-database-create" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.559599 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="0723de91-066f-474b-b7fb-f5f252a830bb" containerName="mariadb-account-create-update" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.559611 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="940d731a-47aa-40db-826b-7c1c7b23ee24" containerName="mariadb-database-create" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.560354 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.576686 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gmp2c-config-vsj5m"] Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.581459 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.664480 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-additional-scripts\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.664557 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-log-ovn\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.664592 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt56b\" (UniqueName: \"kubernetes.io/projected/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-kube-api-access-qt56b\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.664666 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-run-ovn\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.664703 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-run\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.664778 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-scripts\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.767023 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-scripts\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.767166 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-additional-scripts\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.767211 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-log-ovn\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.767678 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-log-ovn\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.767749 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt56b\" (UniqueName: \"kubernetes.io/projected/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-kube-api-access-qt56b\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.768350 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-run-ovn\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.768428 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-run\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.768435 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-run-ovn\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.768343 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-additional-scripts\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.768583 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-run\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.770394 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-scripts\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.789640 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt56b\" (UniqueName: \"kubernetes.io/projected/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-kube-api-access-qt56b\") pod \"ovn-controller-gmp2c-config-vsj5m\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:43 crc kubenswrapper[4872]: I0126 09:22:43.885714 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.339640 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gmp2c-config-vsj5m"] Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.374196 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-k6vff"] Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.378056 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-k6vff" Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.382507 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.384131 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/768400c6-7637-4b2d-b2e3-0123b6fadef0-operator-scripts\") pod \"root-account-create-update-k6vff\" (UID: \"768400c6-7637-4b2d-b2e3-0123b6fadef0\") " pod="openstack/root-account-create-update-k6vff" Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.384186 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vzld\" (UniqueName: \"kubernetes.io/projected/768400c6-7637-4b2d-b2e3-0123b6fadef0-kube-api-access-6vzld\") pod \"root-account-create-update-k6vff\" (UID: \"768400c6-7637-4b2d-b2e3-0123b6fadef0\") " pod="openstack/root-account-create-update-k6vff" Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.404546 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-k6vff"] Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.473082 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"fc8642bd-5341-4e95-be61-deb51d754edb","Type":"ContainerStarted","Data":"6eb358b8ee93d8b4b3da1908bceaddab4b8c4e16862dbc36487aa8a1ec305d4f"} Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.473285 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.475615 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gmp2c-config-vsj5m" event={"ID":"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9","Type":"ContainerStarted","Data":"a353fdd378e190bdba90309623a01e23e3701664e32b136d4a7ee1b91bc43573"} Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.488623 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vzld\" (UniqueName: \"kubernetes.io/projected/768400c6-7637-4b2d-b2e3-0123b6fadef0-kube-api-access-6vzld\") pod \"root-account-create-update-k6vff\" (UID: \"768400c6-7637-4b2d-b2e3-0123b6fadef0\") " pod="openstack/root-account-create-update-k6vff" Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.488785 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/768400c6-7637-4b2d-b2e3-0123b6fadef0-operator-scripts\") pod \"root-account-create-update-k6vff\" (UID: \"768400c6-7637-4b2d-b2e3-0123b6fadef0\") " pod="openstack/root-account-create-update-k6vff" Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.490349 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/768400c6-7637-4b2d-b2e3-0123b6fadef0-operator-scripts\") pod \"root-account-create-update-k6vff\" (UID: \"768400c6-7637-4b2d-b2e3-0123b6fadef0\") " pod="openstack/root-account-create-update-k6vff" Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.509518 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.986067267 podStartE2EDuration="4.509491713s" podCreationTimestamp="2026-01-26 09:22:40 +0000 UTC" firstStartedPulling="2026-01-26 09:22:41.529463296 +0000 UTC m=+894.838303087" lastFinishedPulling="2026-01-26 09:22:43.052887742 +0000 UTC m=+896.361727533" observedRunningTime="2026-01-26 09:22:44.491645203 +0000 UTC m=+897.800485024" watchObservedRunningTime="2026-01-26 09:22:44.509491713 +0000 UTC m=+897.818331524" Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.515907 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vzld\" (UniqueName: \"kubernetes.io/projected/768400c6-7637-4b2d-b2e3-0123b6fadef0-kube-api-access-6vzld\") pod \"root-account-create-update-k6vff\" (UID: \"768400c6-7637-4b2d-b2e3-0123b6fadef0\") " pod="openstack/root-account-create-update-k6vff" Jan 26 09:22:44 crc kubenswrapper[4872]: I0126 09:22:44.749274 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-k6vff" Jan 26 09:22:45 crc kubenswrapper[4872]: I0126 09:22:45.373121 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-k6vff"] Jan 26 09:22:45 crc kubenswrapper[4872]: W0126 09:22:45.378357 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod768400c6_7637_4b2d_b2e3_0123b6fadef0.slice/crio-c2e522a539ed1cfd77bebe0ce0dfbed0f2e6b90455e2ab36efc07fb86fbdd3bd WatchSource:0}: Error finding container c2e522a539ed1cfd77bebe0ce0dfbed0f2e6b90455e2ab36efc07fb86fbdd3bd: Status 404 returned error can't find the container with id c2e522a539ed1cfd77bebe0ce0dfbed0f2e6b90455e2ab36efc07fb86fbdd3bd Jan 26 09:22:45 crc kubenswrapper[4872]: I0126 09:22:45.491241 4872 generic.go:334] "Generic (PLEG): container finished" podID="014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9" containerID="8469ae4b59baa03382c6cc7faa957ae29ebde8d9f27e2a58f0d07160e555b42a" exitCode=0 Jan 26 09:22:45 crc kubenswrapper[4872]: I0126 09:22:45.491397 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gmp2c-config-vsj5m" event={"ID":"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9","Type":"ContainerDied","Data":"8469ae4b59baa03382c6cc7faa957ae29ebde8d9f27e2a58f0d07160e555b42a"} Jan 26 09:22:45 crc kubenswrapper[4872]: I0126 09:22:45.493279 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-k6vff" event={"ID":"768400c6-7637-4b2d-b2e3-0123b6fadef0","Type":"ContainerStarted","Data":"c2e522a539ed1cfd77bebe0ce0dfbed0f2e6b90455e2ab36efc07fb86fbdd3bd"} Jan 26 09:22:45 crc kubenswrapper[4872]: I0126 09:22:45.496954 4872 generic.go:334] "Generic (PLEG): container finished" podID="3ffefe20-1e04-4a87-b2cf-5140aefbb87d" containerID="00b66d2a3b87df557d2d995293ecd78226ba52338e3237259eda5e836c1466a1" exitCode=0 Jan 26 09:22:45 crc kubenswrapper[4872]: I0126 09:22:45.497150 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nvsn5" event={"ID":"3ffefe20-1e04-4a87-b2cf-5140aefbb87d","Type":"ContainerDied","Data":"00b66d2a3b87df557d2d995293ecd78226ba52338e3237259eda5e836c1466a1"} Jan 26 09:22:46 crc kubenswrapper[4872]: I0126 09:22:46.510300 4872 generic.go:334] "Generic (PLEG): container finished" podID="768400c6-7637-4b2d-b2e3-0123b6fadef0" containerID="cbd60b0ea6d59b8ed7de42a25443218b3a80555a9f30aec90412ae7d48fa9ab2" exitCode=0 Jan 26 09:22:46 crc kubenswrapper[4872]: I0126 09:22:46.510426 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-k6vff" event={"ID":"768400c6-7637-4b2d-b2e3-0123b6fadef0","Type":"ContainerDied","Data":"cbd60b0ea6d59b8ed7de42a25443218b3a80555a9f30aec90412ae7d48fa9ab2"} Jan 26 09:22:46 crc kubenswrapper[4872]: I0126 09:22:46.535505 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:46 crc kubenswrapper[4872]: I0126 09:22:46.545381 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f94cf23a-424d-4bd6-8e19-b1f3dc7a0160-etc-swift\") pod \"swift-storage-0\" (UID: \"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160\") " pod="openstack/swift-storage-0" Jan 26 09:22:46 crc kubenswrapper[4872]: I0126 09:22:46.633400 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Jan 26 09:22:46 crc kubenswrapper[4872]: I0126 09:22:46.959214 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.010412 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.045779 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-scripts\") pod \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.045863 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt56b\" (UniqueName: \"kubernetes.io/projected/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-kube-api-access-qt56b\") pod \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.045886 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-swiftconf\") pod \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.045916 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-run-ovn\") pod \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.045941 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-etc-swift\") pod \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.045964 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-scripts\") pod \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.045993 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-run\") pod \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.046018 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-log-ovn\") pod \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.046045 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-combined-ca-bundle\") pod \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.046064 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-additional-scripts\") pod \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\" (UID: \"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.046123 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-ring-data-devices\") pod \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.046144 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-dispersionconf\") pod \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.046179 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pv58\" (UniqueName: \"kubernetes.io/projected/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-kube-api-access-2pv58\") pod \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\" (UID: \"3ffefe20-1e04-4a87-b2cf-5140aefbb87d\") " Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.047393 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-run" (OuterVolumeSpecName: "var-run") pod "014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9" (UID: "014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.047950 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9" (UID: "014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.048337 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9" (UID: "014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.048634 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9" (UID: "014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.049022 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3ffefe20-1e04-4a87-b2cf-5140aefbb87d" (UID: "3ffefe20-1e04-4a87-b2cf-5140aefbb87d"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.049065 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-scripts" (OuterVolumeSpecName: "scripts") pod "014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9" (UID: "014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.049703 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3ffefe20-1e04-4a87-b2cf-5140aefbb87d" (UID: "3ffefe20-1e04-4a87-b2cf-5140aefbb87d"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.054477 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-kube-api-access-qt56b" (OuterVolumeSpecName: "kube-api-access-qt56b") pod "014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9" (UID: "014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9"). InnerVolumeSpecName "kube-api-access-qt56b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.056737 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-kube-api-access-2pv58" (OuterVolumeSpecName: "kube-api-access-2pv58") pod "3ffefe20-1e04-4a87-b2cf-5140aefbb87d" (UID: "3ffefe20-1e04-4a87-b2cf-5140aefbb87d"). InnerVolumeSpecName "kube-api-access-2pv58". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.059664 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3ffefe20-1e04-4a87-b2cf-5140aefbb87d" (UID: "3ffefe20-1e04-4a87-b2cf-5140aefbb87d"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.073816 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-scripts" (OuterVolumeSpecName: "scripts") pod "3ffefe20-1e04-4a87-b2cf-5140aefbb87d" (UID: "3ffefe20-1e04-4a87-b2cf-5140aefbb87d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.075888 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ffefe20-1e04-4a87-b2cf-5140aefbb87d" (UID: "3ffefe20-1e04-4a87-b2cf-5140aefbb87d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.083002 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3ffefe20-1e04-4a87-b2cf-5140aefbb87d" (UID: "3ffefe20-1e04-4a87-b2cf-5140aefbb87d"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150398 4872 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150451 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150469 4872 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150483 4872 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-ring-data-devices\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150495 4872 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-dispersionconf\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150510 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pv58\" (UniqueName: \"kubernetes.io/projected/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-kube-api-access-2pv58\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150524 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150537 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt56b\" (UniqueName: \"kubernetes.io/projected/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-kube-api-access-qt56b\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150553 4872 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-swiftconf\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150567 4872 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150579 4872 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-etc-swift\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150598 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3ffefe20-1e04-4a87-b2cf-5140aefbb87d-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.150618 4872 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9-var-run\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.308994 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.537505 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gmp2c-config-vsj5m" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.537623 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gmp2c-config-vsj5m" event={"ID":"014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9","Type":"ContainerDied","Data":"a353fdd378e190bdba90309623a01e23e3701664e32b136d4a7ee1b91bc43573"} Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.538023 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a353fdd378e190bdba90309623a01e23e3701664e32b136d4a7ee1b91bc43573" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.540337 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"336cabaad467863eb7029323b52647f7c71e240cdbe70255f9c1d930a666985b"} Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.555035 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nvsn5" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.555150 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nvsn5" event={"ID":"3ffefe20-1e04-4a87-b2cf-5140aefbb87d","Type":"ContainerDied","Data":"a5f26b5adc59bc27b8fa79fdb014a6a440f8c752655874eb073762d413c9237e"} Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.555191 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5f26b5adc59bc27b8fa79fdb014a6a440f8c752655874eb073762d413c9237e" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.985687 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-tchz8"] Jan 26 09:22:47 crc kubenswrapper[4872]: E0126 09:22:47.986208 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ffefe20-1e04-4a87-b2cf-5140aefbb87d" containerName="swift-ring-rebalance" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.986228 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ffefe20-1e04-4a87-b2cf-5140aefbb87d" containerName="swift-ring-rebalance" Jan 26 09:22:47 crc kubenswrapper[4872]: E0126 09:22:47.986249 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9" containerName="ovn-config" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.986257 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9" containerName="ovn-config" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.986447 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ffefe20-1e04-4a87-b2cf-5140aefbb87d" containerName="swift-ring-rebalance" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.986473 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9" containerName="ovn-config" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.987163 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.992131 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-n7wxv" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.992429 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Jan 26 09:22:47 crc kubenswrapper[4872]: I0126 09:22:47.999067 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tchz8"] Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.053425 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-k6vff" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.068290 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vzld\" (UniqueName: \"kubernetes.io/projected/768400c6-7637-4b2d-b2e3-0123b6fadef0-kube-api-access-6vzld\") pod \"768400c6-7637-4b2d-b2e3-0123b6fadef0\" (UID: \"768400c6-7637-4b2d-b2e3-0123b6fadef0\") " Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.068484 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/768400c6-7637-4b2d-b2e3-0123b6fadef0-operator-scripts\") pod \"768400c6-7637-4b2d-b2e3-0123b6fadef0\" (UID: \"768400c6-7637-4b2d-b2e3-0123b6fadef0\") " Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.069001 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-combined-ca-bundle\") pod \"glance-db-sync-tchz8\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.069068 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfk24\" (UniqueName: \"kubernetes.io/projected/342ea714-cff5-4b5b-bf68-f7baf40bd84e-kube-api-access-kfk24\") pod \"glance-db-sync-tchz8\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.069110 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-db-sync-config-data\") pod \"glance-db-sync-tchz8\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.069127 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-config-data\") pod \"glance-db-sync-tchz8\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.070042 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/768400c6-7637-4b2d-b2e3-0123b6fadef0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "768400c6-7637-4b2d-b2e3-0123b6fadef0" (UID: "768400c6-7637-4b2d-b2e3-0123b6fadef0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.078275 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/768400c6-7637-4b2d-b2e3-0123b6fadef0-kube-api-access-6vzld" (OuterVolumeSpecName: "kube-api-access-6vzld") pod "768400c6-7637-4b2d-b2e3-0123b6fadef0" (UID: "768400c6-7637-4b2d-b2e3-0123b6fadef0"). InnerVolumeSpecName "kube-api-access-6vzld". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.153741 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gmp2c-config-vsj5m"] Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.162279 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gmp2c-config-vsj5m"] Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.169591 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-gmp2c" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.172141 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfk24\" (UniqueName: \"kubernetes.io/projected/342ea714-cff5-4b5b-bf68-f7baf40bd84e-kube-api-access-kfk24\") pod \"glance-db-sync-tchz8\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.172243 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-db-sync-config-data\") pod \"glance-db-sync-tchz8\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.172273 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-config-data\") pod \"glance-db-sync-tchz8\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.172443 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-combined-ca-bundle\") pod \"glance-db-sync-tchz8\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.172513 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/768400c6-7637-4b2d-b2e3-0123b6fadef0-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.172532 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vzld\" (UniqueName: \"kubernetes.io/projected/768400c6-7637-4b2d-b2e3-0123b6fadef0-kube-api-access-6vzld\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.177096 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-db-sync-config-data\") pod \"glance-db-sync-tchz8\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.179329 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-config-data\") pod \"glance-db-sync-tchz8\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.179557 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-combined-ca-bundle\") pod \"glance-db-sync-tchz8\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.205188 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfk24\" (UniqueName: \"kubernetes.io/projected/342ea714-cff5-4b5b-bf68-f7baf40bd84e-kube-api-access-kfk24\") pod \"glance-db-sync-tchz8\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.210664 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-gmp2c-config-zkvqr"] Jan 26 09:22:48 crc kubenswrapper[4872]: E0126 09:22:48.211274 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="768400c6-7637-4b2d-b2e3-0123b6fadef0" containerName="mariadb-account-create-update" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.211297 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="768400c6-7637-4b2d-b2e3-0123b6fadef0" containerName="mariadb-account-create-update" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.211486 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="768400c6-7637-4b2d-b2e3-0123b6fadef0" containerName="mariadb-account-create-update" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.212243 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.214613 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.272974 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/63b2787b-8cb9-4008-a5eb-fb0e950f156e-additional-scripts\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.273054 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-log-ovn\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.273074 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63b2787b-8cb9-4008-a5eb-fb0e950f156e-scripts\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.273120 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tsfh\" (UniqueName: \"kubernetes.io/projected/63b2787b-8cb9-4008-a5eb-fb0e950f156e-kube-api-access-5tsfh\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.273147 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-run-ovn\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.273193 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-run\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.305757 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gmp2c-config-zkvqr"] Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.372255 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tchz8" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.376478 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/63b2787b-8cb9-4008-a5eb-fb0e950f156e-additional-scripts\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.376628 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-log-ovn\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.376662 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63b2787b-8cb9-4008-a5eb-fb0e950f156e-scripts\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.376712 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tsfh\" (UniqueName: \"kubernetes.io/projected/63b2787b-8cb9-4008-a5eb-fb0e950f156e-kube-api-access-5tsfh\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.376757 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-run-ovn\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.376831 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-run\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.377534 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-run\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.378600 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/63b2787b-8cb9-4008-a5eb-fb0e950f156e-additional-scripts\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.378680 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-log-ovn\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.379412 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-run-ovn\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.381246 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63b2787b-8cb9-4008-a5eb-fb0e950f156e-scripts\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.406897 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tsfh\" (UniqueName: \"kubernetes.io/projected/63b2787b-8cb9-4008-a5eb-fb0e950f156e-kube-api-access-5tsfh\") pod \"ovn-controller-gmp2c-config-zkvqr\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.484004 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.572078 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.573942 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-k6vff" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.580908 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-k6vff" event={"ID":"768400c6-7637-4b2d-b2e3-0123b6fadef0","Type":"ContainerDied","Data":"c2e522a539ed1cfd77bebe0ce0dfbed0f2e6b90455e2ab36efc07fb86fbdd3bd"} Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.580958 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2e522a539ed1cfd77bebe0ce0dfbed0f2e6b90455e2ab36efc07fb86fbdd3bd" Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.601707 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-k8xxm"] Jan 26 09:22:48 crc kubenswrapper[4872]: I0126 09:22:48.602336 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" podUID="130ac93b-f4db-4e8b-9637-f05cda3d1f0c" containerName="dnsmasq-dns" containerID="cri-o://91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec" gracePeriod=10 Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.201743 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9" path="/var/lib/kubelet/pods/014a506e-bcd1-4b1f-94f2-c4ac9c69d1b9/volumes" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.252683 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tchz8"] Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.260878 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-gmp2c-config-zkvqr"] Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.458297 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.505249 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-dns-svc\") pod \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\" (UID: \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\") " Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.505515 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-config\") pod \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\" (UID: \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\") " Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.505619 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttxpx\" (UniqueName: \"kubernetes.io/projected/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-kube-api-access-ttxpx\") pod \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\" (UID: \"130ac93b-f4db-4e8b-9637-f05cda3d1f0c\") " Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.520689 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-kube-api-access-ttxpx" (OuterVolumeSpecName: "kube-api-access-ttxpx") pod "130ac93b-f4db-4e8b-9637-f05cda3d1f0c" (UID: "130ac93b-f4db-4e8b-9637-f05cda3d1f0c"). InnerVolumeSpecName "kube-api-access-ttxpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.552371 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "130ac93b-f4db-4e8b-9637-f05cda3d1f0c" (UID: "130ac93b-f4db-4e8b-9637-f05cda3d1f0c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.553984 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-config" (OuterVolumeSpecName: "config") pod "130ac93b-f4db-4e8b-9637-f05cda3d1f0c" (UID: "130ac93b-f4db-4e8b-9637-f05cda3d1f0c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.584502 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tchz8" event={"ID":"342ea714-cff5-4b5b-bf68-f7baf40bd84e","Type":"ContainerStarted","Data":"487783d4c9f452165ab3b1251ca626f255ec01af279284c1418257b2c41ebac9"} Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.590285 4872 generic.go:334] "Generic (PLEG): container finished" podID="130ac93b-f4db-4e8b-9637-f05cda3d1f0c" containerID="91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec" exitCode=0 Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.590477 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.591916 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" event={"ID":"130ac93b-f4db-4e8b-9637-f05cda3d1f0c","Type":"ContainerDied","Data":"91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec"} Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.591996 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-k8xxm" event={"ID":"130ac93b-f4db-4e8b-9637-f05cda3d1f0c","Type":"ContainerDied","Data":"d737b945610968488bae37de443696b658df73f5111bd538347e7353e1cac604"} Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.592027 4872 scope.go:117] "RemoveContainer" containerID="91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.599688 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gmp2c-config-zkvqr" event={"ID":"63b2787b-8cb9-4008-a5eb-fb0e950f156e","Type":"ContainerStarted","Data":"e1e818850c4279f45b1442869bccd596c2e3d2597e47f235517faefded02132c"} Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.608988 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.609030 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttxpx\" (UniqueName: \"kubernetes.io/projected/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-kube-api-access-ttxpx\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.609041 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/130ac93b-f4db-4e8b-9637-f05cda3d1f0c-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.660055 4872 scope.go:117] "RemoveContainer" containerID="ebf6afebed792a45b43a70bca753d1fdb5f85a3773d2b1bf63dfb0edeb146680" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.668966 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-k8xxm"] Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.698142 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-k8xxm"] Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.723251 4872 scope.go:117] "RemoveContainer" containerID="91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec" Jan 26 09:22:49 crc kubenswrapper[4872]: E0126 09:22:49.725415 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec\": container with ID starting with 91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec not found: ID does not exist" containerID="91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.725460 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec"} err="failed to get container status \"91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec\": rpc error: code = NotFound desc = could not find container \"91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec\": container with ID starting with 91de3abbc296d4c6e8ab418ef76a77be24bc89cf0053a252ff2cf0c4955956ec not found: ID does not exist" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.725498 4872 scope.go:117] "RemoveContainer" containerID="ebf6afebed792a45b43a70bca753d1fdb5f85a3773d2b1bf63dfb0edeb146680" Jan 26 09:22:49 crc kubenswrapper[4872]: E0126 09:22:49.726035 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebf6afebed792a45b43a70bca753d1fdb5f85a3773d2b1bf63dfb0edeb146680\": container with ID starting with ebf6afebed792a45b43a70bca753d1fdb5f85a3773d2b1bf63dfb0edeb146680 not found: ID does not exist" containerID="ebf6afebed792a45b43a70bca753d1fdb5f85a3773d2b1bf63dfb0edeb146680" Jan 26 09:22:49 crc kubenswrapper[4872]: I0126 09:22:49.726093 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebf6afebed792a45b43a70bca753d1fdb5f85a3773d2b1bf63dfb0edeb146680"} err="failed to get container status \"ebf6afebed792a45b43a70bca753d1fdb5f85a3773d2b1bf63dfb0edeb146680\": rpc error: code = NotFound desc = could not find container \"ebf6afebed792a45b43a70bca753d1fdb5f85a3773d2b1bf63dfb0edeb146680\": container with ID starting with ebf6afebed792a45b43a70bca753d1fdb5f85a3773d2b1bf63dfb0edeb146680 not found: ID does not exist" Jan 26 09:22:50 crc kubenswrapper[4872]: I0126 09:22:50.617642 4872 generic.go:334] "Generic (PLEG): container finished" podID="63b2787b-8cb9-4008-a5eb-fb0e950f156e" containerID="cca8635ce74eaf63ddace9a7113b3ad50e65c19034ad7968f82733272bbd826e" exitCode=0 Jan 26 09:22:50 crc kubenswrapper[4872]: I0126 09:22:50.617821 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gmp2c-config-zkvqr" event={"ID":"63b2787b-8cb9-4008-a5eb-fb0e950f156e","Type":"ContainerDied","Data":"cca8635ce74eaf63ddace9a7113b3ad50e65c19034ad7968f82733272bbd826e"} Jan 26 09:22:50 crc kubenswrapper[4872]: I0126 09:22:50.620988 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"77c204031209471627c9de5e5db68b629ba41b34900ca6dc968898bbf9896155"} Jan 26 09:22:50 crc kubenswrapper[4872]: I0126 09:22:50.621043 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"dfb65ba2c13dd259f61aa7c7b322ddc7a9a9b46db2627dff2716faa6b1d8ae32"} Jan 26 09:22:50 crc kubenswrapper[4872]: I0126 09:22:50.621055 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"ead4f84e0927e1958a49d3ff0b42d8554f9a9247aaba16def082e90ac0a4e0ff"} Jan 26 09:22:50 crc kubenswrapper[4872]: I0126 09:22:50.861369 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-k6vff"] Jan 26 09:22:50 crc kubenswrapper[4872]: I0126 09:22:50.871046 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-k6vff"] Jan 26 09:22:51 crc kubenswrapper[4872]: I0126 09:22:51.197641 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="130ac93b-f4db-4e8b-9637-f05cda3d1f0c" path="/var/lib/kubelet/pods/130ac93b-f4db-4e8b-9637-f05cda3d1f0c/volumes" Jan 26 09:22:51 crc kubenswrapper[4872]: I0126 09:22:51.198509 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="768400c6-7637-4b2d-b2e3-0123b6fadef0" path="/var/lib/kubelet/pods/768400c6-7637-4b2d-b2e3-0123b6fadef0/volumes" Jan 26 09:22:51 crc kubenswrapper[4872]: I0126 09:22:51.663383 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"dd51fb9cdb7fe4a2953e283eb7e39a33c2282b3d18b07333a2a8f97aaab52143"} Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.226676 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.376174 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63b2787b-8cb9-4008-a5eb-fb0e950f156e-scripts\") pod \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.376403 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-run-ovn\") pod \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.376472 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-log-ovn\") pod \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.376552 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-run\") pod \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.376612 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tsfh\" (UniqueName: \"kubernetes.io/projected/63b2787b-8cb9-4008-a5eb-fb0e950f156e-kube-api-access-5tsfh\") pod \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.376605 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "63b2787b-8cb9-4008-a5eb-fb0e950f156e" (UID: "63b2787b-8cb9-4008-a5eb-fb0e950f156e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.376672 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/63b2787b-8cb9-4008-a5eb-fb0e950f156e-additional-scripts\") pod \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\" (UID: \"63b2787b-8cb9-4008-a5eb-fb0e950f156e\") " Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.376705 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "63b2787b-8cb9-4008-a5eb-fb0e950f156e" (UID: "63b2787b-8cb9-4008-a5eb-fb0e950f156e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.376692 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-run" (OuterVolumeSpecName: "var-run") pod "63b2787b-8cb9-4008-a5eb-fb0e950f156e" (UID: "63b2787b-8cb9-4008-a5eb-fb0e950f156e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.377139 4872 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-run\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.377614 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63b2787b-8cb9-4008-a5eb-fb0e950f156e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "63b2787b-8cb9-4008-a5eb-fb0e950f156e" (UID: "63b2787b-8cb9-4008-a5eb-fb0e950f156e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.377957 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63b2787b-8cb9-4008-a5eb-fb0e950f156e-scripts" (OuterVolumeSpecName: "scripts") pod "63b2787b-8cb9-4008-a5eb-fb0e950f156e" (UID: "63b2787b-8cb9-4008-a5eb-fb0e950f156e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.378302 4872 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.378344 4872 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/63b2787b-8cb9-4008-a5eb-fb0e950f156e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.381523 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63b2787b-8cb9-4008-a5eb-fb0e950f156e-kube-api-access-5tsfh" (OuterVolumeSpecName: "kube-api-access-5tsfh") pod "63b2787b-8cb9-4008-a5eb-fb0e950f156e" (UID: "63b2787b-8cb9-4008-a5eb-fb0e950f156e"). InnerVolumeSpecName "kube-api-access-5tsfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.480276 4872 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/63b2787b-8cb9-4008-a5eb-fb0e950f156e-additional-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.480321 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/63b2787b-8cb9-4008-a5eb-fb0e950f156e-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.480332 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tsfh\" (UniqueName: \"kubernetes.io/projected/63b2787b-8cb9-4008-a5eb-fb0e950f156e-kube-api-access-5tsfh\") on node \"crc\" DevicePath \"\"" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.678519 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-gmp2c-config-zkvqr" event={"ID":"63b2787b-8cb9-4008-a5eb-fb0e950f156e","Type":"ContainerDied","Data":"e1e818850c4279f45b1442869bccd596c2e3d2597e47f235517faefded02132c"} Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.679117 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1e818850c4279f45b1442869bccd596c2e3d2597e47f235517faefded02132c" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.678639 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-gmp2c-config-zkvqr" Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.693076 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"02acafc5bc19d3d019c90a4abd6f475e3fac81e8b0ea86b89353c8ccc8dd7767"} Jan 26 09:22:52 crc kubenswrapper[4872]: I0126 09:22:52.693152 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"871471f613c079e6a59b68503b847a4244fd59f533227d82bc1edced2576dfc1"} Jan 26 09:22:53 crc kubenswrapper[4872]: I0126 09:22:53.347009 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-gmp2c-config-zkvqr"] Jan 26 09:22:53 crc kubenswrapper[4872]: I0126 09:22:53.413320 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-gmp2c-config-zkvqr"] Jan 26 09:22:53 crc kubenswrapper[4872]: I0126 09:22:53.711829 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"b8252dc67463ca32e1ccb62438390f816e3268890fd6bdaca646527c3a0a2eba"} Jan 26 09:22:53 crc kubenswrapper[4872]: I0126 09:22:53.711894 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"020cd40a31d897b323b28a35d33d0dbb2448278d2f114d05e1dec98e505e3a70"} Jan 26 09:22:54 crc kubenswrapper[4872]: I0126 09:22:54.723015 4872 generic.go:334] "Generic (PLEG): container finished" podID="0792ad43-38af-4c94-bd3b-fc7d3a7af5da" containerID="61e80e60ff53238778a51ec238b504652ca6a6c35b0dc148831642ff362f238a" exitCode=0 Jan 26 09:22:54 crc kubenswrapper[4872]: I0126 09:22:54.723112 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0792ad43-38af-4c94-bd3b-fc7d3a7af5da","Type":"ContainerDied","Data":"61e80e60ff53238778a51ec238b504652ca6a6c35b0dc148831642ff362f238a"} Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.196021 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63b2787b-8cb9-4008-a5eb-fb0e950f156e" path="/var/lib/kubelet/pods/63b2787b-8cb9-4008-a5eb-fb0e950f156e/volumes" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.739411 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0792ad43-38af-4c94-bd3b-fc7d3a7af5da","Type":"ContainerStarted","Data":"2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5"} Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.740127 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.744647 4872 generic.go:334] "Generic (PLEG): container finished" podID="82c8133d-d299-4c99-b8a0-96099e834e8c" containerID="ad7fd4226f3c6d842b398d7962b8022bc46582f511d8ebd06feb9754c684ed0d" exitCode=0 Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.744699 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"82c8133d-d299-4c99-b8a0-96099e834e8c","Type":"ContainerDied","Data":"ad7fd4226f3c6d842b398d7962b8022bc46582f511d8ebd06feb9754c684ed0d"} Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.755495 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"f370ef6f0111db1aaf7e5011b371efce12f14bd8ce5179c9bcb761e49ced400d"} Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.755553 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"3d80be9c00e3b2a9a04039dcf06f794a6d3bc3ce012cfc68222ef38e499992c1"} Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.775478 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.273845745 podStartE2EDuration="1m22.775447368s" podCreationTimestamp="2026-01-26 09:21:33 +0000 UTC" firstStartedPulling="2026-01-26 09:21:35.456070199 +0000 UTC m=+828.764910000" lastFinishedPulling="2026-01-26 09:22:20.957671822 +0000 UTC m=+874.266511623" observedRunningTime="2026-01-26 09:22:55.767323863 +0000 UTC m=+909.076163664" watchObservedRunningTime="2026-01-26 09:22:55.775447368 +0000 UTC m=+909.084287189" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.901420 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-dnrdk"] Jan 26 09:22:55 crc kubenswrapper[4872]: E0126 09:22:55.901774 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63b2787b-8cb9-4008-a5eb-fb0e950f156e" containerName="ovn-config" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.901787 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="63b2787b-8cb9-4008-a5eb-fb0e950f156e" containerName="ovn-config" Jan 26 09:22:55 crc kubenswrapper[4872]: E0126 09:22:55.901808 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130ac93b-f4db-4e8b-9637-f05cda3d1f0c" containerName="init" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.901815 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="130ac93b-f4db-4e8b-9637-f05cda3d1f0c" containerName="init" Jan 26 09:22:55 crc kubenswrapper[4872]: E0126 09:22:55.901828 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130ac93b-f4db-4e8b-9637-f05cda3d1f0c" containerName="dnsmasq-dns" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.901833 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="130ac93b-f4db-4e8b-9637-f05cda3d1f0c" containerName="dnsmasq-dns" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.902007 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="63b2787b-8cb9-4008-a5eb-fb0e950f156e" containerName="ovn-config" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.902027 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="130ac93b-f4db-4e8b-9637-f05cda3d1f0c" containerName="dnsmasq-dns" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.902589 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-dnrdk" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.917628 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01da7c35-0645-45eb-814b-7bfe35d14f46-operator-scripts\") pod \"root-account-create-update-dnrdk\" (UID: \"01da7c35-0645-45eb-814b-7bfe35d14f46\") " pod="openstack/root-account-create-update-dnrdk" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.917704 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzjc8\" (UniqueName: \"kubernetes.io/projected/01da7c35-0645-45eb-814b-7bfe35d14f46-kube-api-access-jzjc8\") pod \"root-account-create-update-dnrdk\" (UID: \"01da7c35-0645-45eb-814b-7bfe35d14f46\") " pod="openstack/root-account-create-update-dnrdk" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.917753 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Jan 26 09:22:55 crc kubenswrapper[4872]: I0126 09:22:55.929766 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-dnrdk"] Jan 26 09:22:56 crc kubenswrapper[4872]: I0126 09:22:56.022402 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01da7c35-0645-45eb-814b-7bfe35d14f46-operator-scripts\") pod \"root-account-create-update-dnrdk\" (UID: \"01da7c35-0645-45eb-814b-7bfe35d14f46\") " pod="openstack/root-account-create-update-dnrdk" Jan 26 09:22:56 crc kubenswrapper[4872]: I0126 09:22:56.022470 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzjc8\" (UniqueName: \"kubernetes.io/projected/01da7c35-0645-45eb-814b-7bfe35d14f46-kube-api-access-jzjc8\") pod \"root-account-create-update-dnrdk\" (UID: \"01da7c35-0645-45eb-814b-7bfe35d14f46\") " pod="openstack/root-account-create-update-dnrdk" Jan 26 09:22:56 crc kubenswrapper[4872]: I0126 09:22:56.029157 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01da7c35-0645-45eb-814b-7bfe35d14f46-operator-scripts\") pod \"root-account-create-update-dnrdk\" (UID: \"01da7c35-0645-45eb-814b-7bfe35d14f46\") " pod="openstack/root-account-create-update-dnrdk" Jan 26 09:22:56 crc kubenswrapper[4872]: I0126 09:22:56.048700 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzjc8\" (UniqueName: \"kubernetes.io/projected/01da7c35-0645-45eb-814b-7bfe35d14f46-kube-api-access-jzjc8\") pod \"root-account-create-update-dnrdk\" (UID: \"01da7c35-0645-45eb-814b-7bfe35d14f46\") " pod="openstack/root-account-create-update-dnrdk" Jan 26 09:22:56 crc kubenswrapper[4872]: I0126 09:22:56.265330 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-dnrdk" Jan 26 09:22:56 crc kubenswrapper[4872]: I0126 09:22:56.761838 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-dnrdk"] Jan 26 09:22:56 crc kubenswrapper[4872]: I0126 09:22:56.767372 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"82c8133d-d299-4c99-b8a0-96099e834e8c","Type":"ContainerStarted","Data":"6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a"} Jan 26 09:22:56 crc kubenswrapper[4872]: I0126 09:22:56.767760 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 26 09:22:56 crc kubenswrapper[4872]: I0126 09:22:56.779519 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"6085843d0fe5c7ed11ec988bb2637e70bc3e3578f8718bf83589bcf79cafef29"} Jan 26 09:22:56 crc kubenswrapper[4872]: I0126 09:22:56.779579 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"39a3df52909373fdbca05dff369767d67c216d662c3c379ca4d891488b7d2cb7"} Jan 26 09:22:56 crc kubenswrapper[4872]: I0126 09:22:56.779600 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"2696a7b521c6ded3d225e15d3151f4d825266b772906034c6cb54924e692c3e4"} Jan 26 09:22:56 crc kubenswrapper[4872]: I0126 09:22:56.812748 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.765528717 podStartE2EDuration="1m24.81272607s" podCreationTimestamp="2026-01-26 09:21:32 +0000 UTC" firstStartedPulling="2026-01-26 09:21:34.938192517 +0000 UTC m=+828.247032318" lastFinishedPulling="2026-01-26 09:22:20.98538987 +0000 UTC m=+874.294229671" observedRunningTime="2026-01-26 09:22:56.807504088 +0000 UTC m=+910.116343889" watchObservedRunningTime="2026-01-26 09:22:56.81272607 +0000 UTC m=+910.121565871" Jan 26 09:22:57 crc kubenswrapper[4872]: I0126 09:22:57.798913 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-dnrdk" event={"ID":"01da7c35-0645-45eb-814b-7bfe35d14f46","Type":"ContainerStarted","Data":"45fbe5ba81b3fe8b1eb6cd55ca2181757f25b189dcf36d10983a72c64806eae0"} Jan 26 09:22:57 crc kubenswrapper[4872]: I0126 09:22:57.808224 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"8f676853004c3a8e167b1023faf1ebdd8c1c6114a15a3f7a91da748f162c6909"} Jan 26 09:23:01 crc kubenswrapper[4872]: I0126 09:23:01.175919 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Jan 26 09:23:07 crc kubenswrapper[4872]: E0126 09:23:07.902037 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Jan 26 09:23:07 crc kubenswrapper[4872]: E0126 09:23:07.902857 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kfk24,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-tchz8_openstack(342ea714-cff5-4b5b-bf68-f7baf40bd84e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:23:07 crc kubenswrapper[4872]: E0126 09:23:07.904095 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-tchz8" podUID="342ea714-cff5-4b5b-bf68-f7baf40bd84e" Jan 26 09:23:07 crc kubenswrapper[4872]: E0126 09:23:07.995873 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-tchz8" podUID="342ea714-cff5-4b5b-bf68-f7baf40bd84e" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.003455 4872 generic.go:334] "Generic (PLEG): container finished" podID="01da7c35-0645-45eb-814b-7bfe35d14f46" containerID="9f1e344a48efae4a82a77a45f54bc9822ea9f569e4d9cd53d0c966e274ec5d9d" exitCode=0 Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.003541 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-dnrdk" event={"ID":"01da7c35-0645-45eb-814b-7bfe35d14f46","Type":"ContainerDied","Data":"9f1e344a48efae4a82a77a45f54bc9822ea9f569e4d9cd53d0c966e274ec5d9d"} Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.012778 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f94cf23a-424d-4bd6-8e19-b1f3dc7a0160","Type":"ContainerStarted","Data":"3cd74c9e33144760028c453923a913257fd2dcebb0f0bd4686d36d3cd6586833"} Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.371787 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=33.071465012 podStartE2EDuration="40.371762858s" podCreationTimestamp="2026-01-26 09:22:29 +0000 UTC" firstStartedPulling="2026-01-26 09:22:47.32628753 +0000 UTC m=+900.635127331" lastFinishedPulling="2026-01-26 09:22:54.626585376 +0000 UTC m=+907.935425177" observedRunningTime="2026-01-26 09:23:09.062484689 +0000 UTC m=+922.371324490" watchObservedRunningTime="2026-01-26 09:23:09.371762858 +0000 UTC m=+922.680602659" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.377686 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-p9gpz"] Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.379700 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.381927 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.394587 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-p9gpz"] Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.552150 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dvjf\" (UniqueName: \"kubernetes.io/projected/1b1367db-cf5f-45c6-b0e6-b9dee6095824-kube-api-access-7dvjf\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.552406 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.552444 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.552733 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.552912 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.553167 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-config\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.655774 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.655912 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.656653 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.656726 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.656894 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-config\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.657002 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dvjf\" (UniqueName: \"kubernetes.io/projected/1b1367db-cf5f-45c6-b0e6-b9dee6095824-kube-api-access-7dvjf\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.657207 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.658105 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.658857 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.659298 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-config\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.659488 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.689073 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dvjf\" (UniqueName: \"kubernetes.io/projected/1b1367db-cf5f-45c6-b0e6-b9dee6095824-kube-api-access-7dvjf\") pod \"dnsmasq-dns-77585f5f8c-p9gpz\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:09 crc kubenswrapper[4872]: I0126 09:23:09.701116 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:10 crc kubenswrapper[4872]: I0126 09:23:10.193358 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-p9gpz"] Jan 26 09:23:10 crc kubenswrapper[4872]: I0126 09:23:10.227549 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:23:10 crc kubenswrapper[4872]: I0126 09:23:10.227615 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:23:10 crc kubenswrapper[4872]: I0126 09:23:10.296004 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-dnrdk" Jan 26 09:23:10 crc kubenswrapper[4872]: I0126 09:23:10.475777 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzjc8\" (UniqueName: \"kubernetes.io/projected/01da7c35-0645-45eb-814b-7bfe35d14f46-kube-api-access-jzjc8\") pod \"01da7c35-0645-45eb-814b-7bfe35d14f46\" (UID: \"01da7c35-0645-45eb-814b-7bfe35d14f46\") " Jan 26 09:23:10 crc kubenswrapper[4872]: I0126 09:23:10.476500 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01da7c35-0645-45eb-814b-7bfe35d14f46-operator-scripts\") pod \"01da7c35-0645-45eb-814b-7bfe35d14f46\" (UID: \"01da7c35-0645-45eb-814b-7bfe35d14f46\") " Jan 26 09:23:10 crc kubenswrapper[4872]: I0126 09:23:10.476982 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01da7c35-0645-45eb-814b-7bfe35d14f46-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "01da7c35-0645-45eb-814b-7bfe35d14f46" (UID: "01da7c35-0645-45eb-814b-7bfe35d14f46"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:10 crc kubenswrapper[4872]: I0126 09:23:10.480788 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01da7c35-0645-45eb-814b-7bfe35d14f46-kube-api-access-jzjc8" (OuterVolumeSpecName: "kube-api-access-jzjc8") pod "01da7c35-0645-45eb-814b-7bfe35d14f46" (UID: "01da7c35-0645-45eb-814b-7bfe35d14f46"). InnerVolumeSpecName "kube-api-access-jzjc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:10 crc kubenswrapper[4872]: I0126 09:23:10.579016 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/01da7c35-0645-45eb-814b-7bfe35d14f46-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:10 crc kubenswrapper[4872]: I0126 09:23:10.579063 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzjc8\" (UniqueName: \"kubernetes.io/projected/01da7c35-0645-45eb-814b-7bfe35d14f46-kube-api-access-jzjc8\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:11 crc kubenswrapper[4872]: I0126 09:23:11.032036 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-dnrdk" Jan 26 09:23:11 crc kubenswrapper[4872]: I0126 09:23:11.033015 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-dnrdk" event={"ID":"01da7c35-0645-45eb-814b-7bfe35d14f46","Type":"ContainerDied","Data":"45fbe5ba81b3fe8b1eb6cd55ca2181757f25b189dcf36d10983a72c64806eae0"} Jan 26 09:23:11 crc kubenswrapper[4872]: I0126 09:23:11.033057 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45fbe5ba81b3fe8b1eb6cd55ca2181757f25b189dcf36d10983a72c64806eae0" Jan 26 09:23:11 crc kubenswrapper[4872]: I0126 09:23:11.035293 4872 generic.go:334] "Generic (PLEG): container finished" podID="1b1367db-cf5f-45c6-b0e6-b9dee6095824" containerID="fa361760c26fdc8a9fca1802f8e52984ede98605e464c94607388227ac364b11" exitCode=0 Jan 26 09:23:11 crc kubenswrapper[4872]: I0126 09:23:11.035325 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" event={"ID":"1b1367db-cf5f-45c6-b0e6-b9dee6095824","Type":"ContainerDied","Data":"fa361760c26fdc8a9fca1802f8e52984ede98605e464c94607388227ac364b11"} Jan 26 09:23:11 crc kubenswrapper[4872]: I0126 09:23:11.035359 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" event={"ID":"1b1367db-cf5f-45c6-b0e6-b9dee6095824","Type":"ContainerStarted","Data":"def2322ba18e2adaadc8ab0504e828186e6bfe0b3fc9667587a1a712698127c5"} Jan 26 09:23:12 crc kubenswrapper[4872]: I0126 09:23:12.049347 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" event={"ID":"1b1367db-cf5f-45c6-b0e6-b9dee6095824","Type":"ContainerStarted","Data":"fd481a6bd06ffca5a8b856db10d95109dd0a586ba26be602f072a9d9c42c1a10"} Jan 26 09:23:12 crc kubenswrapper[4872]: I0126 09:23:12.050028 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:12 crc kubenswrapper[4872]: I0126 09:23:12.075541 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" podStartSLOduration=3.075514837 podStartE2EDuration="3.075514837s" podCreationTimestamp="2026-01-26 09:23:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:23:12.069336421 +0000 UTC m=+925.378176242" watchObservedRunningTime="2026-01-26 09:23:12.075514837 +0000 UTC m=+925.384354638" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.278163 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.598113 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-v4kh2"] Jan 26 09:23:14 crc kubenswrapper[4872]: E0126 09:23:14.598544 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01da7c35-0645-45eb-814b-7bfe35d14f46" containerName="mariadb-account-create-update" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.598564 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="01da7c35-0645-45eb-814b-7bfe35d14f46" containerName="mariadb-account-create-update" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.598719 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="01da7c35-0645-45eb-814b-7bfe35d14f46" containerName="mariadb-account-create-update" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.599352 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v4kh2" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.615996 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.624647 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-v4kh2"] Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.710496 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-24ac-account-create-update-vp5nc"] Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.711637 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-24ac-account-create-update-vp5nc" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.722713 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-28cqb"] Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.723960 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-28cqb" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.732204 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.741651 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-24ac-account-create-update-vp5nc"] Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.753069 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grgzf\" (UniqueName: \"kubernetes.io/projected/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d-kube-api-access-grgzf\") pod \"cinder-db-create-v4kh2\" (UID: \"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d\") " pod="openstack/cinder-db-create-v4kh2" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.753171 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d-operator-scripts\") pod \"cinder-db-create-v4kh2\" (UID: \"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d\") " pod="openstack/cinder-db-create-v4kh2" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.759431 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-28cqb"] Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.813222 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-0c9f-account-create-update-q688k"] Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.817892 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0c9f-account-create-update-q688k" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.827387 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.854457 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-0c9f-account-create-update-q688k"] Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.855705 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2a972ca-1c4f-4124-b481-0eeca0e1ac06-operator-scripts\") pod \"barbican-24ac-account-create-update-vp5nc\" (UID: \"d2a972ca-1c4f-4124-b481-0eeca0e1ac06\") " pod="openstack/barbican-24ac-account-create-update-vp5nc" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.855793 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hr27\" (UniqueName: \"kubernetes.io/projected/d2a972ca-1c4f-4124-b481-0eeca0e1ac06-kube-api-access-6hr27\") pod \"barbican-24ac-account-create-update-vp5nc\" (UID: \"d2a972ca-1c4f-4124-b481-0eeca0e1ac06\") " pod="openstack/barbican-24ac-account-create-update-vp5nc" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.855849 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc7c1498-7864-484d-b86a-2f9ba512bf13-operator-scripts\") pod \"cinder-0c9f-account-create-update-q688k\" (UID: \"cc7c1498-7864-484d-b86a-2f9ba512bf13\") " pod="openstack/cinder-0c9f-account-create-update-q688k" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.855911 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r89vh\" (UniqueName: \"kubernetes.io/projected/4d5889bf-aa3d-4588-9d97-4c184bca015b-kube-api-access-r89vh\") pod \"barbican-db-create-28cqb\" (UID: \"4d5889bf-aa3d-4588-9d97-4c184bca015b\") " pod="openstack/barbican-db-create-28cqb" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.855959 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grgzf\" (UniqueName: \"kubernetes.io/projected/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d-kube-api-access-grgzf\") pod \"cinder-db-create-v4kh2\" (UID: \"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d\") " pod="openstack/cinder-db-create-v4kh2" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.856031 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6plz\" (UniqueName: \"kubernetes.io/projected/cc7c1498-7864-484d-b86a-2f9ba512bf13-kube-api-access-j6plz\") pod \"cinder-0c9f-account-create-update-q688k\" (UID: \"cc7c1498-7864-484d-b86a-2f9ba512bf13\") " pod="openstack/cinder-0c9f-account-create-update-q688k" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.856069 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d5889bf-aa3d-4588-9d97-4c184bca015b-operator-scripts\") pod \"barbican-db-create-28cqb\" (UID: \"4d5889bf-aa3d-4588-9d97-4c184bca015b\") " pod="openstack/barbican-db-create-28cqb" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.856106 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d-operator-scripts\") pod \"cinder-db-create-v4kh2\" (UID: \"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d\") " pod="openstack/cinder-db-create-v4kh2" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.857221 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d-operator-scripts\") pod \"cinder-db-create-v4kh2\" (UID: \"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d\") " pod="openstack/cinder-db-create-v4kh2" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.887158 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grgzf\" (UniqueName: \"kubernetes.io/projected/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d-kube-api-access-grgzf\") pod \"cinder-db-create-v4kh2\" (UID: \"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d\") " pod="openstack/cinder-db-create-v4kh2" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.925084 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v4kh2" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.957395 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r89vh\" (UniqueName: \"kubernetes.io/projected/4d5889bf-aa3d-4588-9d97-4c184bca015b-kube-api-access-r89vh\") pod \"barbican-db-create-28cqb\" (UID: \"4d5889bf-aa3d-4588-9d97-4c184bca015b\") " pod="openstack/barbican-db-create-28cqb" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.957476 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6plz\" (UniqueName: \"kubernetes.io/projected/cc7c1498-7864-484d-b86a-2f9ba512bf13-kube-api-access-j6plz\") pod \"cinder-0c9f-account-create-update-q688k\" (UID: \"cc7c1498-7864-484d-b86a-2f9ba512bf13\") " pod="openstack/cinder-0c9f-account-create-update-q688k" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.957511 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d5889bf-aa3d-4588-9d97-4c184bca015b-operator-scripts\") pod \"barbican-db-create-28cqb\" (UID: \"4d5889bf-aa3d-4588-9d97-4c184bca015b\") " pod="openstack/barbican-db-create-28cqb" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.957590 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2a972ca-1c4f-4124-b481-0eeca0e1ac06-operator-scripts\") pod \"barbican-24ac-account-create-update-vp5nc\" (UID: \"d2a972ca-1c4f-4124-b481-0eeca0e1ac06\") " pod="openstack/barbican-24ac-account-create-update-vp5nc" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.957608 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hr27\" (UniqueName: \"kubernetes.io/projected/d2a972ca-1c4f-4124-b481-0eeca0e1ac06-kube-api-access-6hr27\") pod \"barbican-24ac-account-create-update-vp5nc\" (UID: \"d2a972ca-1c4f-4124-b481-0eeca0e1ac06\") " pod="openstack/barbican-24ac-account-create-update-vp5nc" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.957631 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc7c1498-7864-484d-b86a-2f9ba512bf13-operator-scripts\") pod \"cinder-0c9f-account-create-update-q688k\" (UID: \"cc7c1498-7864-484d-b86a-2f9ba512bf13\") " pod="openstack/cinder-0c9f-account-create-update-q688k" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.958566 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc7c1498-7864-484d-b86a-2f9ba512bf13-operator-scripts\") pod \"cinder-0c9f-account-create-update-q688k\" (UID: \"cc7c1498-7864-484d-b86a-2f9ba512bf13\") " pod="openstack/cinder-0c9f-account-create-update-q688k" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.959203 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2a972ca-1c4f-4124-b481-0eeca0e1ac06-operator-scripts\") pod \"barbican-24ac-account-create-update-vp5nc\" (UID: \"d2a972ca-1c4f-4124-b481-0eeca0e1ac06\") " pod="openstack/barbican-24ac-account-create-update-vp5nc" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.959288 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d5889bf-aa3d-4588-9d97-4c184bca015b-operator-scripts\") pod \"barbican-db-create-28cqb\" (UID: \"4d5889bf-aa3d-4588-9d97-4c184bca015b\") " pod="openstack/barbican-db-create-28cqb" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.979576 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-ccwl7"] Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.984534 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.995156 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.995449 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.996080 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.996323 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-7k5wq" Jan 26 09:23:14 crc kubenswrapper[4872]: I0126 09:23:14.996896 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ccwl7"] Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.000865 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6plz\" (UniqueName: \"kubernetes.io/projected/cc7c1498-7864-484d-b86a-2f9ba512bf13-kube-api-access-j6plz\") pod \"cinder-0c9f-account-create-update-q688k\" (UID: \"cc7c1498-7864-484d-b86a-2f9ba512bf13\") " pod="openstack/cinder-0c9f-account-create-update-q688k" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.008487 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r89vh\" (UniqueName: \"kubernetes.io/projected/4d5889bf-aa3d-4588-9d97-4c184bca015b-kube-api-access-r89vh\") pod \"barbican-db-create-28cqb\" (UID: \"4d5889bf-aa3d-4588-9d97-4c184bca015b\") " pod="openstack/barbican-db-create-28cqb" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.016447 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hr27\" (UniqueName: \"kubernetes.io/projected/d2a972ca-1c4f-4124-b481-0eeca0e1ac06-kube-api-access-6hr27\") pod \"barbican-24ac-account-create-update-vp5nc\" (UID: \"d2a972ca-1c4f-4124-b481-0eeca0e1ac06\") " pod="openstack/barbican-24ac-account-create-update-vp5nc" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.019217 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-xf6vh"] Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.020873 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xf6vh" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.029545 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-24ac-account-create-update-vp5nc" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.036850 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xf6vh"] Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.077842 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxh76\" (UniqueName: \"kubernetes.io/projected/8759676c-42c7-41ef-b71e-f9848dd84ceb-kube-api-access-qxh76\") pod \"neutron-db-create-xf6vh\" (UID: \"8759676c-42c7-41ef-b71e-f9848dd84ceb\") " pod="openstack/neutron-db-create-xf6vh" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.077949 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8759676c-42c7-41ef-b71e-f9848dd84ceb-operator-scripts\") pod \"neutron-db-create-xf6vh\" (UID: \"8759676c-42c7-41ef-b71e-f9848dd84ceb\") " pod="openstack/neutron-db-create-xf6vh" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.077983 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrvxc\" (UniqueName: \"kubernetes.io/projected/b3741809-0b4f-4816-9130-c9f8d73aaded-kube-api-access-mrvxc\") pod \"keystone-db-sync-ccwl7\" (UID: \"b3741809-0b4f-4816-9130-c9f8d73aaded\") " pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.078109 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3741809-0b4f-4816-9130-c9f8d73aaded-combined-ca-bundle\") pod \"keystone-db-sync-ccwl7\" (UID: \"b3741809-0b4f-4816-9130-c9f8d73aaded\") " pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.078168 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3741809-0b4f-4816-9130-c9f8d73aaded-config-data\") pod \"keystone-db-sync-ccwl7\" (UID: \"b3741809-0b4f-4816-9130-c9f8d73aaded\") " pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.078667 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-28cqb" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.128146 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6421-account-create-update-75kn6"] Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.130085 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6421-account-create-update-75kn6" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.141055 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.147843 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6421-account-create-update-75kn6"] Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.157586 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0c9f-account-create-update-q688k" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.179193 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20add17d-dd15-47ff-8d0c-5c6ca089fc65-operator-scripts\") pod \"neutron-6421-account-create-update-75kn6\" (UID: \"20add17d-dd15-47ff-8d0c-5c6ca089fc65\") " pod="openstack/neutron-6421-account-create-update-75kn6" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.179246 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3741809-0b4f-4816-9130-c9f8d73aaded-config-data\") pod \"keystone-db-sync-ccwl7\" (UID: \"b3741809-0b4f-4816-9130-c9f8d73aaded\") " pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.179289 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxh76\" (UniqueName: \"kubernetes.io/projected/8759676c-42c7-41ef-b71e-f9848dd84ceb-kube-api-access-qxh76\") pod \"neutron-db-create-xf6vh\" (UID: \"8759676c-42c7-41ef-b71e-f9848dd84ceb\") " pod="openstack/neutron-db-create-xf6vh" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.179331 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8759676c-42c7-41ef-b71e-f9848dd84ceb-operator-scripts\") pod \"neutron-db-create-xf6vh\" (UID: \"8759676c-42c7-41ef-b71e-f9848dd84ceb\") " pod="openstack/neutron-db-create-xf6vh" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.179348 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrvxc\" (UniqueName: \"kubernetes.io/projected/b3741809-0b4f-4816-9130-c9f8d73aaded-kube-api-access-mrvxc\") pod \"keystone-db-sync-ccwl7\" (UID: \"b3741809-0b4f-4816-9130-c9f8d73aaded\") " pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.179412 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd9qp\" (UniqueName: \"kubernetes.io/projected/20add17d-dd15-47ff-8d0c-5c6ca089fc65-kube-api-access-dd9qp\") pod \"neutron-6421-account-create-update-75kn6\" (UID: \"20add17d-dd15-47ff-8d0c-5c6ca089fc65\") " pod="openstack/neutron-6421-account-create-update-75kn6" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.179433 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3741809-0b4f-4816-9130-c9f8d73aaded-combined-ca-bundle\") pod \"keystone-db-sync-ccwl7\" (UID: \"b3741809-0b4f-4816-9130-c9f8d73aaded\") " pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.185814 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3741809-0b4f-4816-9130-c9f8d73aaded-combined-ca-bundle\") pod \"keystone-db-sync-ccwl7\" (UID: \"b3741809-0b4f-4816-9130-c9f8d73aaded\") " pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.189567 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3741809-0b4f-4816-9130-c9f8d73aaded-config-data\") pod \"keystone-db-sync-ccwl7\" (UID: \"b3741809-0b4f-4816-9130-c9f8d73aaded\") " pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.190905 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8759676c-42c7-41ef-b71e-f9848dd84ceb-operator-scripts\") pod \"neutron-db-create-xf6vh\" (UID: \"8759676c-42c7-41ef-b71e-f9848dd84ceb\") " pod="openstack/neutron-db-create-xf6vh" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.214556 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxh76\" (UniqueName: \"kubernetes.io/projected/8759676c-42c7-41ef-b71e-f9848dd84ceb-kube-api-access-qxh76\") pod \"neutron-db-create-xf6vh\" (UID: \"8759676c-42c7-41ef-b71e-f9848dd84ceb\") " pod="openstack/neutron-db-create-xf6vh" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.221790 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrvxc\" (UniqueName: \"kubernetes.io/projected/b3741809-0b4f-4816-9130-c9f8d73aaded-kube-api-access-mrvxc\") pod \"keystone-db-sync-ccwl7\" (UID: \"b3741809-0b4f-4816-9130-c9f8d73aaded\") " pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.222567 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xf6vh" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.282304 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd9qp\" (UniqueName: \"kubernetes.io/projected/20add17d-dd15-47ff-8d0c-5c6ca089fc65-kube-api-access-dd9qp\") pod \"neutron-6421-account-create-update-75kn6\" (UID: \"20add17d-dd15-47ff-8d0c-5c6ca089fc65\") " pod="openstack/neutron-6421-account-create-update-75kn6" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.282374 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20add17d-dd15-47ff-8d0c-5c6ca089fc65-operator-scripts\") pod \"neutron-6421-account-create-update-75kn6\" (UID: \"20add17d-dd15-47ff-8d0c-5c6ca089fc65\") " pod="openstack/neutron-6421-account-create-update-75kn6" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.284211 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20add17d-dd15-47ff-8d0c-5c6ca089fc65-operator-scripts\") pod \"neutron-6421-account-create-update-75kn6\" (UID: \"20add17d-dd15-47ff-8d0c-5c6ca089fc65\") " pod="openstack/neutron-6421-account-create-update-75kn6" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.306594 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd9qp\" (UniqueName: \"kubernetes.io/projected/20add17d-dd15-47ff-8d0c-5c6ca089fc65-kube-api-access-dd9qp\") pod \"neutron-6421-account-create-update-75kn6\" (UID: \"20add17d-dd15-47ff-8d0c-5c6ca089fc65\") " pod="openstack/neutron-6421-account-create-update-75kn6" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.473625 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.535706 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6421-account-create-update-75kn6" Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.675551 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-v4kh2"] Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.781327 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-24ac-account-create-update-vp5nc"] Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.843766 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-28cqb"] Jan 26 09:23:15 crc kubenswrapper[4872]: W0126 09:23:15.855096 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d5889bf_aa3d_4588_9d97_4c184bca015b.slice/crio-e9c0ebf31b2f7d82e31428ece91615f120f39992dae531bcfafcac7466c414c8 WatchSource:0}: Error finding container e9c0ebf31b2f7d82e31428ece91615f120f39992dae531bcfafcac7466c414c8: Status 404 returned error can't find the container with id e9c0ebf31b2f7d82e31428ece91615f120f39992dae531bcfafcac7466c414c8 Jan 26 09:23:15 crc kubenswrapper[4872]: I0126 09:23:15.996487 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-0c9f-account-create-update-q688k"] Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.031255 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-xf6vh"] Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.062982 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6421-account-create-update-75kn6"] Jan 26 09:23:16 crc kubenswrapper[4872]: W0126 09:23:16.081898 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20add17d_dd15_47ff_8d0c_5c6ca089fc65.slice/crio-dfac2747a44a1d61635112b0b236762dd90a2f0a7160915529b6953f547408c2 WatchSource:0}: Error finding container dfac2747a44a1d61635112b0b236762dd90a2f0a7160915529b6953f547408c2: Status 404 returned error can't find the container with id dfac2747a44a1d61635112b0b236762dd90a2f0a7160915529b6953f547408c2 Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.124753 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ccwl7"] Jan 26 09:23:16 crc kubenswrapper[4872]: W0126 09:23:16.130069 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3741809_0b4f_4816_9130_c9f8d73aaded.slice/crio-6fb58a655cca98d9f678eda96606812258cff6bd7ccbe789be63119f67402397 WatchSource:0}: Error finding container 6fb58a655cca98d9f678eda96606812258cff6bd7ccbe789be63119f67402397: Status 404 returned error can't find the container with id 6fb58a655cca98d9f678eda96606812258cff6bd7ccbe789be63119f67402397 Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.143083 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-24ac-account-create-update-vp5nc" event={"ID":"d2a972ca-1c4f-4124-b481-0eeca0e1ac06","Type":"ContainerStarted","Data":"32b2ef75c3bf80927edfec877a66d09ed6e86666394fdb8b5858bf3dae6e5205"} Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.143163 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-24ac-account-create-update-vp5nc" event={"ID":"d2a972ca-1c4f-4124-b481-0eeca0e1ac06","Type":"ContainerStarted","Data":"207d85ce52f8e3c356bc34cb2c0a4c64a3205e1eba9ccb10f85eed62c1d3795a"} Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.146934 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0c9f-account-create-update-q688k" event={"ID":"cc7c1498-7864-484d-b86a-2f9ba512bf13","Type":"ContainerStarted","Data":"d93480dc235b7e1a15c9b52d773fa67a5a3d07198040c2b0fa968d943f237bd0"} Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.149208 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-28cqb" event={"ID":"4d5889bf-aa3d-4588-9d97-4c184bca015b","Type":"ContainerStarted","Data":"e9c0ebf31b2f7d82e31428ece91615f120f39992dae531bcfafcac7466c414c8"} Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.150538 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v4kh2" event={"ID":"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d","Type":"ContainerStarted","Data":"8c0611984b881bce81eac16dd33b546dba9189a4544cd647d0da16b009a27346"} Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.150577 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v4kh2" event={"ID":"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d","Type":"ContainerStarted","Data":"79003437ea7d09c4ac346a2348b50d986cb663ec3e53e6c95c7ad543401355c3"} Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.155401 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6421-account-create-update-75kn6" event={"ID":"20add17d-dd15-47ff-8d0c-5c6ca089fc65","Type":"ContainerStarted","Data":"dfac2747a44a1d61635112b0b236762dd90a2f0a7160915529b6953f547408c2"} Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.167165 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xf6vh" event={"ID":"8759676c-42c7-41ef-b71e-f9848dd84ceb","Type":"ContainerStarted","Data":"30d6ee9591a22ec68da04e9c0003b59255cf63a78e3ba8ea609d306cad65a0df"} Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.172852 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-24ac-account-create-update-vp5nc" podStartSLOduration=2.172778599 podStartE2EDuration="2.172778599s" podCreationTimestamp="2026-01-26 09:23:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:23:16.164718566 +0000 UTC m=+929.473558387" watchObservedRunningTime="2026-01-26 09:23:16.172778599 +0000 UTC m=+929.481618400" Jan 26 09:23:16 crc kubenswrapper[4872]: I0126 09:23:16.198812 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-v4kh2" podStartSLOduration=2.198767674 podStartE2EDuration="2.198767674s" podCreationTimestamp="2026-01-26 09:23:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:23:16.187617353 +0000 UTC m=+929.496457174" watchObservedRunningTime="2026-01-26 09:23:16.198767674 +0000 UTC m=+929.507607475" Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.180822 4872 generic.go:334] "Generic (PLEG): container finished" podID="20add17d-dd15-47ff-8d0c-5c6ca089fc65" containerID="b27def704e2ed5900d0a9e45a5bff4f0c1e906e0031f7970dd6c733e48aaa678" exitCode=0 Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.182321 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6421-account-create-update-75kn6" event={"ID":"20add17d-dd15-47ff-8d0c-5c6ca089fc65","Type":"ContainerDied","Data":"b27def704e2ed5900d0a9e45a5bff4f0c1e906e0031f7970dd6c733e48aaa678"} Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.185034 4872 generic.go:334] "Generic (PLEG): container finished" podID="8759676c-42c7-41ef-b71e-f9848dd84ceb" containerID="7f97ad1a95a364f7c129725296d2cdac6e3572e58dc16246927a099f72bf3f66" exitCode=0 Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.188287 4872 generic.go:334] "Generic (PLEG): container finished" podID="d2a972ca-1c4f-4124-b481-0eeca0e1ac06" containerID="32b2ef75c3bf80927edfec877a66d09ed6e86666394fdb8b5858bf3dae6e5205" exitCode=0 Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.192333 4872 generic.go:334] "Generic (PLEG): container finished" podID="cc7c1498-7864-484d-b86a-2f9ba512bf13" containerID="285157d6fe4d1fdddbc90cb5c3d52d6ea2d827a0cecf5ac04d4964503cedc339" exitCode=0 Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.195254 4872 generic.go:334] "Generic (PLEG): container finished" podID="4d5889bf-aa3d-4588-9d97-4c184bca015b" containerID="f827292a70bfdbade694480b6f9973315117fbf091e98aa6a6ab496f2aaa0b15" exitCode=0 Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.199960 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xf6vh" event={"ID":"8759676c-42c7-41ef-b71e-f9848dd84ceb","Type":"ContainerDied","Data":"7f97ad1a95a364f7c129725296d2cdac6e3572e58dc16246927a099f72bf3f66"} Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.200013 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-24ac-account-create-update-vp5nc" event={"ID":"d2a972ca-1c4f-4124-b481-0eeca0e1ac06","Type":"ContainerDied","Data":"32b2ef75c3bf80927edfec877a66d09ed6e86666394fdb8b5858bf3dae6e5205"} Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.200029 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ccwl7" event={"ID":"b3741809-0b4f-4816-9130-c9f8d73aaded","Type":"ContainerStarted","Data":"6fb58a655cca98d9f678eda96606812258cff6bd7ccbe789be63119f67402397"} Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.200040 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0c9f-account-create-update-q688k" event={"ID":"cc7c1498-7864-484d-b86a-2f9ba512bf13","Type":"ContainerDied","Data":"285157d6fe4d1fdddbc90cb5c3d52d6ea2d827a0cecf5ac04d4964503cedc339"} Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.200054 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-28cqb" event={"ID":"4d5889bf-aa3d-4588-9d97-4c184bca015b","Type":"ContainerDied","Data":"f827292a70bfdbade694480b6f9973315117fbf091e98aa6a6ab496f2aaa0b15"} Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.202680 4872 generic.go:334] "Generic (PLEG): container finished" podID="096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d" containerID="8c0611984b881bce81eac16dd33b546dba9189a4544cd647d0da16b009a27346" exitCode=0 Jan 26 09:23:17 crc kubenswrapper[4872]: I0126 09:23:17.202910 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v4kh2" event={"ID":"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d","Type":"ContainerDied","Data":"8c0611984b881bce81eac16dd33b546dba9189a4544cd647d0da16b009a27346"} Jan 26 09:23:18 crc kubenswrapper[4872]: I0126 09:23:18.810562 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v4kh2" Jan 26 09:23:18 crc kubenswrapper[4872]: I0126 09:23:18.859296 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grgzf\" (UniqueName: \"kubernetes.io/projected/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d-kube-api-access-grgzf\") pod \"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d\" (UID: \"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d\") " Jan 26 09:23:18 crc kubenswrapper[4872]: I0126 09:23:18.859384 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d-operator-scripts\") pod \"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d\" (UID: \"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d\") " Jan 26 09:23:18 crc kubenswrapper[4872]: I0126 09:23:18.861076 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d" (UID: "096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:18 crc kubenswrapper[4872]: I0126 09:23:18.873757 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d-kube-api-access-grgzf" (OuterVolumeSpecName: "kube-api-access-grgzf") pod "096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d" (UID: "096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d"). InnerVolumeSpecName "kube-api-access-grgzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:18 crc kubenswrapper[4872]: I0126 09:23:18.961979 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grgzf\" (UniqueName: \"kubernetes.io/projected/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d-kube-api-access-grgzf\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:18 crc kubenswrapper[4872]: I0126 09:23:18.962036 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:19 crc kubenswrapper[4872]: I0126 09:23:19.221937 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-v4kh2" event={"ID":"096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d","Type":"ContainerDied","Data":"79003437ea7d09c4ac346a2348b50d986cb663ec3e53e6c95c7ad543401355c3"} Jan 26 09:23:19 crc kubenswrapper[4872]: I0126 09:23:19.221983 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79003437ea7d09c4ac346a2348b50d986cb663ec3e53e6c95c7ad543401355c3" Jan 26 09:23:19 crc kubenswrapper[4872]: I0126 09:23:19.222046 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-v4kh2" Jan 26 09:23:19 crc kubenswrapper[4872]: I0126 09:23:19.701940 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:23:19 crc kubenswrapper[4872]: I0126 09:23:19.817819 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qdxbg"] Jan 26 09:23:19 crc kubenswrapper[4872]: I0126 09:23:19.818164 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-qdxbg" podUID="9ba3969e-0cea-4a89-8092-b2cf63786ecf" containerName="dnsmasq-dns" containerID="cri-o://14a8d88b6802e3b097ff366b7cd26b9edda8faaa71177069aef9dc53ca1ce533" gracePeriod=10 Jan 26 09:23:20 crc kubenswrapper[4872]: I0126 09:23:20.235034 4872 generic.go:334] "Generic (PLEG): container finished" podID="9ba3969e-0cea-4a89-8092-b2cf63786ecf" containerID="14a8d88b6802e3b097ff366b7cd26b9edda8faaa71177069aef9dc53ca1ce533" exitCode=0 Jan 26 09:23:20 crc kubenswrapper[4872]: I0126 09:23:20.235577 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qdxbg" event={"ID":"9ba3969e-0cea-4a89-8092-b2cf63786ecf","Type":"ContainerDied","Data":"14a8d88b6802e3b097ff366b7cd26b9edda8faaa71177069aef9dc53ca1ce533"} Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.869888 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6421-account-create-update-75kn6" Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.912610 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-28cqb" Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.914837 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-24ac-account-create-update-vp5nc" Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.929201 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xf6vh" Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.929685 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20add17d-dd15-47ff-8d0c-5c6ca089fc65-operator-scripts\") pod \"20add17d-dd15-47ff-8d0c-5c6ca089fc65\" (UID: \"20add17d-dd15-47ff-8d0c-5c6ca089fc65\") " Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.929935 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r89vh\" (UniqueName: \"kubernetes.io/projected/4d5889bf-aa3d-4588-9d97-4c184bca015b-kube-api-access-r89vh\") pod \"4d5889bf-aa3d-4588-9d97-4c184bca015b\" (UID: \"4d5889bf-aa3d-4588-9d97-4c184bca015b\") " Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.929988 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dd9qp\" (UniqueName: \"kubernetes.io/projected/20add17d-dd15-47ff-8d0c-5c6ca089fc65-kube-api-access-dd9qp\") pod \"20add17d-dd15-47ff-8d0c-5c6ca089fc65\" (UID: \"20add17d-dd15-47ff-8d0c-5c6ca089fc65\") " Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.930025 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d5889bf-aa3d-4588-9d97-4c184bca015b-operator-scripts\") pod \"4d5889bf-aa3d-4588-9d97-4c184bca015b\" (UID: \"4d5889bf-aa3d-4588-9d97-4c184bca015b\") " Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.932599 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d5889bf-aa3d-4588-9d97-4c184bca015b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4d5889bf-aa3d-4588-9d97-4c184bca015b" (UID: "4d5889bf-aa3d-4588-9d97-4c184bca015b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.933096 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20add17d-dd15-47ff-8d0c-5c6ca089fc65-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "20add17d-dd15-47ff-8d0c-5c6ca089fc65" (UID: "20add17d-dd15-47ff-8d0c-5c6ca089fc65"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.937574 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d5889bf-aa3d-4588-9d97-4c184bca015b-kube-api-access-r89vh" (OuterVolumeSpecName: "kube-api-access-r89vh") pod "4d5889bf-aa3d-4588-9d97-4c184bca015b" (UID: "4d5889bf-aa3d-4588-9d97-4c184bca015b"). InnerVolumeSpecName "kube-api-access-r89vh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.940200 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0c9f-account-create-update-q688k" Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.947209 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20add17d-dd15-47ff-8d0c-5c6ca089fc65-kube-api-access-dd9qp" (OuterVolumeSpecName: "kube-api-access-dd9qp") pod "20add17d-dd15-47ff-8d0c-5c6ca089fc65" (UID: "20add17d-dd15-47ff-8d0c-5c6ca089fc65"). InnerVolumeSpecName "kube-api-access-dd9qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:21 crc kubenswrapper[4872]: I0126 09:23:21.970780 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.032299 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-ovsdbserver-nb\") pod \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.032391 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-ovsdbserver-sb\") pod \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.032430 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc7c1498-7864-484d-b86a-2f9ba512bf13-operator-scripts\") pod \"cc7c1498-7864-484d-b86a-2f9ba512bf13\" (UID: \"cc7c1498-7864-484d-b86a-2f9ba512bf13\") " Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.032496 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2a972ca-1c4f-4124-b481-0eeca0e1ac06-operator-scripts\") pod \"d2a972ca-1c4f-4124-b481-0eeca0e1ac06\" (UID: \"d2a972ca-1c4f-4124-b481-0eeca0e1ac06\") " Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.032542 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-config\") pod \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.032601 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-dns-svc\") pod \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.032631 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxh76\" (UniqueName: \"kubernetes.io/projected/8759676c-42c7-41ef-b71e-f9848dd84ceb-kube-api-access-qxh76\") pod \"8759676c-42c7-41ef-b71e-f9848dd84ceb\" (UID: \"8759676c-42c7-41ef-b71e-f9848dd84ceb\") " Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.032726 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t96pc\" (UniqueName: \"kubernetes.io/projected/9ba3969e-0cea-4a89-8092-b2cf63786ecf-kube-api-access-t96pc\") pod \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\" (UID: \"9ba3969e-0cea-4a89-8092-b2cf63786ecf\") " Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.032788 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hr27\" (UniqueName: \"kubernetes.io/projected/d2a972ca-1c4f-4124-b481-0eeca0e1ac06-kube-api-access-6hr27\") pod \"d2a972ca-1c4f-4124-b481-0eeca0e1ac06\" (UID: \"d2a972ca-1c4f-4124-b481-0eeca0e1ac06\") " Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.032818 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6plz\" (UniqueName: \"kubernetes.io/projected/cc7c1498-7864-484d-b86a-2f9ba512bf13-kube-api-access-j6plz\") pod \"cc7c1498-7864-484d-b86a-2f9ba512bf13\" (UID: \"cc7c1498-7864-484d-b86a-2f9ba512bf13\") " Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.032842 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8759676c-42c7-41ef-b71e-f9848dd84ceb-operator-scripts\") pod \"8759676c-42c7-41ef-b71e-f9848dd84ceb\" (UID: \"8759676c-42c7-41ef-b71e-f9848dd84ceb\") " Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.033285 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r89vh\" (UniqueName: \"kubernetes.io/projected/4d5889bf-aa3d-4588-9d97-4c184bca015b-kube-api-access-r89vh\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.033309 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dd9qp\" (UniqueName: \"kubernetes.io/projected/20add17d-dd15-47ff-8d0c-5c6ca089fc65-kube-api-access-dd9qp\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.033320 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d5889bf-aa3d-4588-9d97-4c184bca015b-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.033331 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/20add17d-dd15-47ff-8d0c-5c6ca089fc65-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.034329 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8759676c-42c7-41ef-b71e-f9848dd84ceb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8759676c-42c7-41ef-b71e-f9848dd84ceb" (UID: "8759676c-42c7-41ef-b71e-f9848dd84ceb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.036933 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc7c1498-7864-484d-b86a-2f9ba512bf13-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cc7c1498-7864-484d-b86a-2f9ba512bf13" (UID: "cc7c1498-7864-484d-b86a-2f9ba512bf13"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.038697 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ba3969e-0cea-4a89-8092-b2cf63786ecf-kube-api-access-t96pc" (OuterVolumeSpecName: "kube-api-access-t96pc") pod "9ba3969e-0cea-4a89-8092-b2cf63786ecf" (UID: "9ba3969e-0cea-4a89-8092-b2cf63786ecf"). InnerVolumeSpecName "kube-api-access-t96pc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.039575 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2a972ca-1c4f-4124-b481-0eeca0e1ac06-kube-api-access-6hr27" (OuterVolumeSpecName: "kube-api-access-6hr27") pod "d2a972ca-1c4f-4124-b481-0eeca0e1ac06" (UID: "d2a972ca-1c4f-4124-b481-0eeca0e1ac06"). InnerVolumeSpecName "kube-api-access-6hr27". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.041197 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc7c1498-7864-484d-b86a-2f9ba512bf13-kube-api-access-j6plz" (OuterVolumeSpecName: "kube-api-access-j6plz") pod "cc7c1498-7864-484d-b86a-2f9ba512bf13" (UID: "cc7c1498-7864-484d-b86a-2f9ba512bf13"). InnerVolumeSpecName "kube-api-access-j6plz". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.041312 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2a972ca-1c4f-4124-b481-0eeca0e1ac06-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d2a972ca-1c4f-4124-b481-0eeca0e1ac06" (UID: "d2a972ca-1c4f-4124-b481-0eeca0e1ac06"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.059462 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8759676c-42c7-41ef-b71e-f9848dd84ceb-kube-api-access-qxh76" (OuterVolumeSpecName: "kube-api-access-qxh76") pod "8759676c-42c7-41ef-b71e-f9848dd84ceb" (UID: "8759676c-42c7-41ef-b71e-f9848dd84ceb"). InnerVolumeSpecName "kube-api-access-qxh76". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.094170 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-config" (OuterVolumeSpecName: "config") pod "9ba3969e-0cea-4a89-8092-b2cf63786ecf" (UID: "9ba3969e-0cea-4a89-8092-b2cf63786ecf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.097791 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9ba3969e-0cea-4a89-8092-b2cf63786ecf" (UID: "9ba3969e-0cea-4a89-8092-b2cf63786ecf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.100464 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9ba3969e-0cea-4a89-8092-b2cf63786ecf" (UID: "9ba3969e-0cea-4a89-8092-b2cf63786ecf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.109208 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9ba3969e-0cea-4a89-8092-b2cf63786ecf" (UID: "9ba3969e-0cea-4a89-8092-b2cf63786ecf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.135336 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hr27\" (UniqueName: \"kubernetes.io/projected/d2a972ca-1c4f-4124-b481-0eeca0e1ac06-kube-api-access-6hr27\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.135381 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6plz\" (UniqueName: \"kubernetes.io/projected/cc7c1498-7864-484d-b86a-2f9ba512bf13-kube-api-access-j6plz\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.135393 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8759676c-42c7-41ef-b71e-f9848dd84ceb-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.135402 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.135412 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.135425 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc7c1498-7864-484d-b86a-2f9ba512bf13-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.135435 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d2a972ca-1c4f-4124-b481-0eeca0e1ac06-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.135448 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.135458 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ba3969e-0cea-4a89-8092-b2cf63786ecf-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.135469 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxh76\" (UniqueName: \"kubernetes.io/projected/8759676c-42c7-41ef-b71e-f9848dd84ceb-kube-api-access-qxh76\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.135481 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t96pc\" (UniqueName: \"kubernetes.io/projected/9ba3969e-0cea-4a89-8092-b2cf63786ecf-kube-api-access-t96pc\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.274263 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-28cqb" event={"ID":"4d5889bf-aa3d-4588-9d97-4c184bca015b","Type":"ContainerDied","Data":"e9c0ebf31b2f7d82e31428ece91615f120f39992dae531bcfafcac7466c414c8"} Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.274322 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9c0ebf31b2f7d82e31428ece91615f120f39992dae531bcfafcac7466c414c8" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.274291 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-28cqb" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.278528 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6421-account-create-update-75kn6" event={"ID":"20add17d-dd15-47ff-8d0c-5c6ca089fc65","Type":"ContainerDied","Data":"dfac2747a44a1d61635112b0b236762dd90a2f0a7160915529b6953f547408c2"} Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.278557 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfac2747a44a1d61635112b0b236762dd90a2f0a7160915529b6953f547408c2" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.278648 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6421-account-create-update-75kn6" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.280297 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-qdxbg" event={"ID":"9ba3969e-0cea-4a89-8092-b2cf63786ecf","Type":"ContainerDied","Data":"b6ddbf56dbf9c1716878ae51b75d3b11b9f41bbb9da617ef936fa09bce6536c2"} Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.280358 4872 scope.go:117] "RemoveContainer" containerID="14a8d88b6802e3b097ff366b7cd26b9edda8faaa71177069aef9dc53ca1ce533" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.280525 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-qdxbg" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.287125 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-xf6vh" event={"ID":"8759676c-42c7-41ef-b71e-f9848dd84ceb","Type":"ContainerDied","Data":"30d6ee9591a22ec68da04e9c0003b59255cf63a78e3ba8ea609d306cad65a0df"} Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.287179 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30d6ee9591a22ec68da04e9c0003b59255cf63a78e3ba8ea609d306cad65a0df" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.287272 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-xf6vh" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.303258 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-24ac-account-create-update-vp5nc" event={"ID":"d2a972ca-1c4f-4124-b481-0eeca0e1ac06","Type":"ContainerDied","Data":"207d85ce52f8e3c356bc34cb2c0a4c64a3205e1eba9ccb10f85eed62c1d3795a"} Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.303320 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="207d85ce52f8e3c356bc34cb2c0a4c64a3205e1eba9ccb10f85eed62c1d3795a" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.303468 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-24ac-account-create-update-vp5nc" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.310884 4872 scope.go:117] "RemoveContainer" containerID="708bdc6a21183df94d75a797e55f60f737d9cd47ec1c48e5347d0d550f485a62" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.312497 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ccwl7" event={"ID":"b3741809-0b4f-4816-9130-c9f8d73aaded","Type":"ContainerStarted","Data":"eddd753e23eec8e2b29820b3ed6a10bea1f8da1dc5b84bf2954290cbb853e838"} Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.334369 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-0c9f-account-create-update-q688k" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.334383 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-0c9f-account-create-update-q688k" event={"ID":"cc7c1498-7864-484d-b86a-2f9ba512bf13","Type":"ContainerDied","Data":"d93480dc235b7e1a15c9b52d773fa67a5a3d07198040c2b0fa968d943f237bd0"} Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.334435 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d93480dc235b7e1a15c9b52d773fa67a5a3d07198040c2b0fa968d943f237bd0" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.359936 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-ccwl7" podStartSLOduration=2.800477847 podStartE2EDuration="8.359911151s" podCreationTimestamp="2026-01-26 09:23:14 +0000 UTC" firstStartedPulling="2026-01-26 09:23:16.142736923 +0000 UTC m=+929.451576724" lastFinishedPulling="2026-01-26 09:23:21.702170227 +0000 UTC m=+935.011010028" observedRunningTime="2026-01-26 09:23:22.348241588 +0000 UTC m=+935.657081389" watchObservedRunningTime="2026-01-26 09:23:22.359911151 +0000 UTC m=+935.668750952" Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.384103 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qdxbg"] Jan 26 09:23:22 crc kubenswrapper[4872]: I0126 09:23:22.391201 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-qdxbg"] Jan 26 09:23:23 crc kubenswrapper[4872]: I0126 09:23:23.194463 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ba3969e-0cea-4a89-8092-b2cf63786ecf" path="/var/lib/kubelet/pods/9ba3969e-0cea-4a89-8092-b2cf63786ecf/volumes" Jan 26 09:23:23 crc kubenswrapper[4872]: I0126 09:23:23.346758 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tchz8" event={"ID":"342ea714-cff5-4b5b-bf68-f7baf40bd84e","Type":"ContainerStarted","Data":"ad4e458c514b7df57ac246f939d7104cfd24233db841179b21dd81517401bbc0"} Jan 26 09:23:23 crc kubenswrapper[4872]: I0126 09:23:23.368239 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-tchz8" podStartSLOduration=4.111682468 podStartE2EDuration="36.368219324s" podCreationTimestamp="2026-01-26 09:22:47 +0000 UTC" firstStartedPulling="2026-01-26 09:22:49.437006844 +0000 UTC m=+902.745846645" lastFinishedPulling="2026-01-26 09:23:21.6935437 +0000 UTC m=+935.002383501" observedRunningTime="2026-01-26 09:23:23.365494065 +0000 UTC m=+936.674333856" watchObservedRunningTime="2026-01-26 09:23:23.368219324 +0000 UTC m=+936.677059125" Jan 26 09:23:26 crc kubenswrapper[4872]: I0126 09:23:26.393097 4872 generic.go:334] "Generic (PLEG): container finished" podID="b3741809-0b4f-4816-9130-c9f8d73aaded" containerID="eddd753e23eec8e2b29820b3ed6a10bea1f8da1dc5b84bf2954290cbb853e838" exitCode=0 Jan 26 09:23:26 crc kubenswrapper[4872]: I0126 09:23:26.393224 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ccwl7" event={"ID":"b3741809-0b4f-4816-9130-c9f8d73aaded","Type":"ContainerDied","Data":"eddd753e23eec8e2b29820b3ed6a10bea1f8da1dc5b84bf2954290cbb853e838"} Jan 26 09:23:27 crc kubenswrapper[4872]: I0126 09:23:27.749056 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:27 crc kubenswrapper[4872]: I0126 09:23:27.870978 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3741809-0b4f-4816-9130-c9f8d73aaded-combined-ca-bundle\") pod \"b3741809-0b4f-4816-9130-c9f8d73aaded\" (UID: \"b3741809-0b4f-4816-9130-c9f8d73aaded\") " Jan 26 09:23:27 crc kubenswrapper[4872]: I0126 09:23:27.871106 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3741809-0b4f-4816-9130-c9f8d73aaded-config-data\") pod \"b3741809-0b4f-4816-9130-c9f8d73aaded\" (UID: \"b3741809-0b4f-4816-9130-c9f8d73aaded\") " Jan 26 09:23:27 crc kubenswrapper[4872]: I0126 09:23:27.871168 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrvxc\" (UniqueName: \"kubernetes.io/projected/b3741809-0b4f-4816-9130-c9f8d73aaded-kube-api-access-mrvxc\") pod \"b3741809-0b4f-4816-9130-c9f8d73aaded\" (UID: \"b3741809-0b4f-4816-9130-c9f8d73aaded\") " Jan 26 09:23:27 crc kubenswrapper[4872]: I0126 09:23:27.878688 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3741809-0b4f-4816-9130-c9f8d73aaded-kube-api-access-mrvxc" (OuterVolumeSpecName: "kube-api-access-mrvxc") pod "b3741809-0b4f-4816-9130-c9f8d73aaded" (UID: "b3741809-0b4f-4816-9130-c9f8d73aaded"). InnerVolumeSpecName "kube-api-access-mrvxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:27 crc kubenswrapper[4872]: I0126 09:23:27.901916 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3741809-0b4f-4816-9130-c9f8d73aaded-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b3741809-0b4f-4816-9130-c9f8d73aaded" (UID: "b3741809-0b4f-4816-9130-c9f8d73aaded"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:27 crc kubenswrapper[4872]: I0126 09:23:27.921902 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3741809-0b4f-4816-9130-c9f8d73aaded-config-data" (OuterVolumeSpecName: "config-data") pod "b3741809-0b4f-4816-9130-c9f8d73aaded" (UID: "b3741809-0b4f-4816-9130-c9f8d73aaded"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:27 crc kubenswrapper[4872]: I0126 09:23:27.972889 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3741809-0b4f-4816-9130-c9f8d73aaded-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:27 crc kubenswrapper[4872]: I0126 09:23:27.972936 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3741809-0b4f-4816-9130-c9f8d73aaded-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:27 crc kubenswrapper[4872]: I0126 09:23:27.972948 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrvxc\" (UniqueName: \"kubernetes.io/projected/b3741809-0b4f-4816-9130-c9f8d73aaded-kube-api-access-mrvxc\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.412616 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ccwl7" event={"ID":"b3741809-0b4f-4816-9130-c9f8d73aaded","Type":"ContainerDied","Data":"6fb58a655cca98d9f678eda96606812258cff6bd7ccbe789be63119f67402397"} Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.412661 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fb58a655cca98d9f678eda96606812258cff6bd7ccbe789be63119f67402397" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.412665 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ccwl7" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.719304 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-m695h"] Jan 26 09:23:28 crc kubenswrapper[4872]: E0126 09:23:28.720254 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba3969e-0cea-4a89-8092-b2cf63786ecf" containerName="init" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720277 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba3969e-0cea-4a89-8092-b2cf63786ecf" containerName="init" Jan 26 09:23:28 crc kubenswrapper[4872]: E0126 09:23:28.720295 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20add17d-dd15-47ff-8d0c-5c6ca089fc65" containerName="mariadb-account-create-update" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720303 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="20add17d-dd15-47ff-8d0c-5c6ca089fc65" containerName="mariadb-account-create-update" Jan 26 09:23:28 crc kubenswrapper[4872]: E0126 09:23:28.720322 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2a972ca-1c4f-4124-b481-0eeca0e1ac06" containerName="mariadb-account-create-update" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720330 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2a972ca-1c4f-4124-b481-0eeca0e1ac06" containerName="mariadb-account-create-update" Jan 26 09:23:28 crc kubenswrapper[4872]: E0126 09:23:28.720345 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d5889bf-aa3d-4588-9d97-4c184bca015b" containerName="mariadb-database-create" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720355 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d5889bf-aa3d-4588-9d97-4c184bca015b" containerName="mariadb-database-create" Jan 26 09:23:28 crc kubenswrapper[4872]: E0126 09:23:28.720383 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7c1498-7864-484d-b86a-2f9ba512bf13" containerName="mariadb-account-create-update" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720390 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7c1498-7864-484d-b86a-2f9ba512bf13" containerName="mariadb-account-create-update" Jan 26 09:23:28 crc kubenswrapper[4872]: E0126 09:23:28.720400 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8759676c-42c7-41ef-b71e-f9848dd84ceb" containerName="mariadb-database-create" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720407 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8759676c-42c7-41ef-b71e-f9848dd84ceb" containerName="mariadb-database-create" Jan 26 09:23:28 crc kubenswrapper[4872]: E0126 09:23:28.720419 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3741809-0b4f-4816-9130-c9f8d73aaded" containerName="keystone-db-sync" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720429 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3741809-0b4f-4816-9130-c9f8d73aaded" containerName="keystone-db-sync" Jan 26 09:23:28 crc kubenswrapper[4872]: E0126 09:23:28.720447 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba3969e-0cea-4a89-8092-b2cf63786ecf" containerName="dnsmasq-dns" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720455 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba3969e-0cea-4a89-8092-b2cf63786ecf" containerName="dnsmasq-dns" Jan 26 09:23:28 crc kubenswrapper[4872]: E0126 09:23:28.720470 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d" containerName="mariadb-database-create" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720478 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d" containerName="mariadb-database-create" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720669 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3741809-0b4f-4816-9130-c9f8d73aaded" containerName="keystone-db-sync" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720684 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2a972ca-1c4f-4124-b481-0eeca0e1ac06" containerName="mariadb-account-create-update" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720698 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="8759676c-42c7-41ef-b71e-f9848dd84ceb" containerName="mariadb-database-create" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720713 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d" containerName="mariadb-database-create" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720721 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="20add17d-dd15-47ff-8d0c-5c6ca089fc65" containerName="mariadb-account-create-update" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720734 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ba3969e-0cea-4a89-8092-b2cf63786ecf" containerName="dnsmasq-dns" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720742 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7c1498-7864-484d-b86a-2f9ba512bf13" containerName="mariadb-account-create-update" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.720756 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d5889bf-aa3d-4588-9d97-4c184bca015b" containerName="mariadb-database-create" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.723561 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.748906 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-dhw6k"] Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.750200 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.758754 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.759045 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.759114 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-7k5wq" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.759440 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.759546 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.772715 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-m695h"] Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.786445 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.786542 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.786574 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-dns-svc\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.786626 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-459cw\" (UniqueName: \"kubernetes.io/projected/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-kube-api-access-459cw\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.786655 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-config\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.786685 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.793638 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dhw6k"] Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.889974 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.890046 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-dns-svc\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.890079 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-combined-ca-bundle\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.890127 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-459cw\" (UniqueName: \"kubernetes.io/projected/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-kube-api-access-459cw\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.890162 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-config\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.890180 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-scripts\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.890200 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-fernet-keys\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.890218 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-config-data\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.890246 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.890267 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-credential-keys\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.890291 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.890324 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mrqg\" (UniqueName: \"kubernetes.io/projected/1bdce24d-9f71-495b-8995-32365acdb90e-kube-api-access-8mrqg\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.891253 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-config\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.891369 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.891602 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-dns-svc\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.892080 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.892219 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.928081 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-459cw\" (UniqueName: \"kubernetes.io/projected/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-kube-api-access-459cw\") pod \"dnsmasq-dns-55fff446b9-m695h\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.937113 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-db-create-jdmsk"] Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.938242 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-jdmsk" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.945967 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-create-jdmsk"] Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.995220 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-scripts\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.995271 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-fernet-keys\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.995294 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-config-data\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.995353 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de9df0f9-9456-4bfd-98f7-065a933bb80e-operator-scripts\") pod \"ironic-db-create-jdmsk\" (UID: \"de9df0f9-9456-4bfd-98f7-065a933bb80e\") " pod="openstack/ironic-db-create-jdmsk" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.995385 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-credential-keys\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.995458 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mrqg\" (UniqueName: \"kubernetes.io/projected/1bdce24d-9f71-495b-8995-32365acdb90e-kube-api-access-8mrqg\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.995543 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-combined-ca-bundle\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:28 crc kubenswrapper[4872]: I0126 09:23:28.995599 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr6xv\" (UniqueName: \"kubernetes.io/projected/de9df0f9-9456-4bfd-98f7-065a933bb80e-kube-api-access-kr6xv\") pod \"ironic-db-create-jdmsk\" (UID: \"de9df0f9-9456-4bfd-98f7-065a933bb80e\") " pod="openstack/ironic-db-create-jdmsk" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.011598 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-credential-keys\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.014212 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-combined-ca-bundle\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.016654 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-fernet-keys\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.020558 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-config-data\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.029144 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-scripts\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.053713 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.057874 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-zd82w"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.064494 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zd82w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.070773 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.071089 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.071457 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mrqg\" (UniqueName: \"kubernetes.io/projected/1bdce24d-9f71-495b-8995-32365acdb90e-kube-api-access-8mrqg\") pod \"keystone-bootstrap-dhw6k\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.078452 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.098374 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de9df0f9-9456-4bfd-98f7-065a933bb80e-operator-scripts\") pod \"ironic-db-create-jdmsk\" (UID: \"de9df0f9-9456-4bfd-98f7-065a933bb80e\") " pod="openstack/ironic-db-create-jdmsk" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.098511 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr6xv\" (UniqueName: \"kubernetes.io/projected/de9df0f9-9456-4bfd-98f7-065a933bb80e-kube-api-access-kr6xv\") pod \"ironic-db-create-jdmsk\" (UID: \"de9df0f9-9456-4bfd-98f7-065a933bb80e\") " pod="openstack/ironic-db-create-jdmsk" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.100459 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de9df0f9-9456-4bfd-98f7-065a933bb80e-operator-scripts\") pod \"ironic-db-create-jdmsk\" (UID: \"de9df0f9-9456-4bfd-98f7-065a933bb80e\") " pod="openstack/ironic-db-create-jdmsk" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.104623 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nm7rl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.153056 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-zd82w"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.200537 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8tv7\" (UniqueName: \"kubernetes.io/projected/04076889-57e8-41ad-b47e-6d8e5e23223f-kube-api-access-v8tv7\") pod \"neutron-db-sync-zd82w\" (UID: \"04076889-57e8-41ad-b47e-6d8e5e23223f\") " pod="openstack/neutron-db-sync-zd82w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.200630 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04076889-57e8-41ad-b47e-6d8e5e23223f-combined-ca-bundle\") pod \"neutron-db-sync-zd82w\" (UID: \"04076889-57e8-41ad-b47e-6d8e5e23223f\") " pod="openstack/neutron-db-sync-zd82w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.200775 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/04076889-57e8-41ad-b47e-6d8e5e23223f-config\") pod \"neutron-db-sync-zd82w\" (UID: \"04076889-57e8-41ad-b47e-6d8e5e23223f\") " pod="openstack/neutron-db-sync-zd82w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.207186 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr6xv\" (UniqueName: \"kubernetes.io/projected/de9df0f9-9456-4bfd-98f7-065a933bb80e-kube-api-access-kr6xv\") pod \"ironic-db-create-jdmsk\" (UID: \"de9df0f9-9456-4bfd-98f7-065a933bb80e\") " pod="openstack/ironic-db-create-jdmsk" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.293969 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-c3e0-account-create-update-pfn6b"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.296769 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-4cbzl"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.298046 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-c3e0-account-create-update-pfn6b"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.298070 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4cbzl"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.298192 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.299023 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-c3e0-account-create-update-pfn6b" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.313057 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-db-secret" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.313400 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-gqlrs" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.320042 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.326351 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.327840 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-jdmsk" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.330510 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8tv7\" (UniqueName: \"kubernetes.io/projected/04076889-57e8-41ad-b47e-6d8e5e23223f-kube-api-access-v8tv7\") pod \"neutron-db-sync-zd82w\" (UID: \"04076889-57e8-41ad-b47e-6d8e5e23223f\") " pod="openstack/neutron-db-sync-zd82w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.330594 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04076889-57e8-41ad-b47e-6d8e5e23223f-combined-ca-bundle\") pod \"neutron-db-sync-zd82w\" (UID: \"04076889-57e8-41ad-b47e-6d8e5e23223f\") " pod="openstack/neutron-db-sync-zd82w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.330719 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/04076889-57e8-41ad-b47e-6d8e5e23223f-config\") pod \"neutron-db-sync-zd82w\" (UID: \"04076889-57e8-41ad-b47e-6d8e5e23223f\") " pod="openstack/neutron-db-sync-zd82w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.338663 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04076889-57e8-41ad-b47e-6d8e5e23223f-combined-ca-bundle\") pod \"neutron-db-sync-zd82w\" (UID: \"04076889-57e8-41ad-b47e-6d8e5e23223f\") " pod="openstack/neutron-db-sync-zd82w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.346789 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/04076889-57e8-41ad-b47e-6d8e5e23223f-config\") pod \"neutron-db-sync-zd82w\" (UID: \"04076889-57e8-41ad-b47e-6d8e5e23223f\") " pod="openstack/neutron-db-sync-zd82w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.379363 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.386713 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8tv7\" (UniqueName: \"kubernetes.io/projected/04076889-57e8-41ad-b47e-6d8e5e23223f-kube-api-access-v8tv7\") pod \"neutron-db-sync-zd82w\" (UID: \"04076889-57e8-41ad-b47e-6d8e5e23223f\") " pod="openstack/neutron-db-sync-zd82w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.391980 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.423339 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.423626 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439079 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-log-httpd\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439197 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-db-sync-config-data\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439227 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-combined-ca-bundle\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439267 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n4rm\" (UniqueName: \"kubernetes.io/projected/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00-kube-api-access-4n4rm\") pod \"ironic-c3e0-account-create-update-pfn6b\" (UID: \"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00\") " pod="openstack/ironic-c3e0-account-create-update-pfn6b" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439301 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439330 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvmw7\" (UniqueName: \"kubernetes.io/projected/3f653549-36ce-4db8-9c67-ec95b628e867-kube-api-access-kvmw7\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439397 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-scripts\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439419 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-run-httpd\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439443 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439535 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-config-data\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439565 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-scripts\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439599 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq65j\" (UniqueName: \"kubernetes.io/projected/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-kube-api-access-qq65j\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439635 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-config-data\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439685 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f653549-36ce-4db8-9c67-ec95b628e867-etc-machine-id\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.439722 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00-operator-scripts\") pod \"ironic-c3e0-account-create-update-pfn6b\" (UID: \"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00\") " pod="openstack/ironic-c3e0-account-create-update-pfn6b" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.461641 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zd82w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.472310 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-zdqwd"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.473525 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.479682 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.479788 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.480062 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-x6c24" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.565788 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-db-sync-config-data\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.565910 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-combined-ca-bundle\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.565949 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n4rm\" (UniqueName: \"kubernetes.io/projected/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00-kube-api-access-4n4rm\") pod \"ironic-c3e0-account-create-update-pfn6b\" (UID: \"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00\") " pod="openstack/ironic-c3e0-account-create-update-pfn6b" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.565972 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.565995 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvmw7\" (UniqueName: \"kubernetes.io/projected/3f653549-36ce-4db8-9c67-ec95b628e867-kube-api-access-kvmw7\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566034 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ltww\" (UniqueName: \"kubernetes.io/projected/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-kube-api-access-5ltww\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566050 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-scripts\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566067 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-run-httpd\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566085 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566130 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-config-data\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566147 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-scripts\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566165 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-scripts\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566183 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq65j\" (UniqueName: \"kubernetes.io/projected/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-kube-api-access-qq65j\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566203 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-config-data\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566229 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-config-data\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566256 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f653549-36ce-4db8-9c67-ec95b628e867-etc-machine-id\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566277 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-combined-ca-bundle\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566299 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00-operator-scripts\") pod \"ironic-c3e0-account-create-update-pfn6b\" (UID: \"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00\") " pod="openstack/ironic-c3e0-account-create-update-pfn6b" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566337 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-log-httpd\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.566355 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-logs\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.579523 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-run-httpd\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.580726 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-combined-ca-bundle\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.580776 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f653549-36ce-4db8-9c67-ec95b628e867-etc-machine-id\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.581309 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00-operator-scripts\") pod \"ironic-c3e0-account-create-update-pfn6b\" (UID: \"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00\") " pod="openstack/ironic-c3e0-account-create-update-pfn6b" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.581512 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-log-httpd\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.589399 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.598919 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-config-data\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.599625 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.604884 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-config-data\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.606254 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-scripts\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.618489 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-db-sync-config-data\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.618884 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.622096 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-scripts\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.623883 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n4rm\" (UniqueName: \"kubernetes.io/projected/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00-kube-api-access-4n4rm\") pod \"ironic-c3e0-account-create-update-pfn6b\" (UID: \"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00\") " pod="openstack/ironic-c3e0-account-create-update-pfn6b" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.625469 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq65j\" (UniqueName: \"kubernetes.io/projected/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-kube-api-access-qq65j\") pod \"ceilometer-0\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.627636 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zdqwd"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.635254 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvmw7\" (UniqueName: \"kubernetes.io/projected/3f653549-36ce-4db8-9c67-ec95b628e867-kube-api-access-kvmw7\") pod \"cinder-db-sync-4cbzl\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.641814 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-m695h"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.676178 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-combined-ca-bundle\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.676300 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-logs\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.676487 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ltww\" (UniqueName: \"kubernetes.io/projected/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-kube-api-access-5ltww\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.676645 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-scripts\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.676677 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-config-data\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.678355 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-logs\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.681912 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-config-data\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.682003 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-dk64w"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.683506 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dk64w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.683557 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-scripts\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.686101 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-zrmml" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.686394 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.686831 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.696635 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dk64w"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.699878 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-c3e0-account-create-update-pfn6b" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.705612 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ltww\" (UniqueName: \"kubernetes.io/projected/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-kube-api-access-5ltww\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.712651 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-combined-ca-bundle\") pod \"placement-db-sync-zdqwd\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.773661 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-xqnrf"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.775160 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.779227 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnslt\" (UniqueName: \"kubernetes.io/projected/ddd0e60e-a07c-41b0-8c44-00844dfef24a-kube-api-access-vnslt\") pod \"barbican-db-sync-dk64w\" (UID: \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\") " pod="openstack/barbican-db-sync-dk64w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.779320 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd0e60e-a07c-41b0-8c44-00844dfef24a-combined-ca-bundle\") pod \"barbican-db-sync-dk64w\" (UID: \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\") " pod="openstack/barbican-db-sync-dk64w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.779381 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ddd0e60e-a07c-41b0-8c44-00844dfef24a-db-sync-config-data\") pod \"barbican-db-sync-dk64w\" (UID: \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\") " pod="openstack/barbican-db-sync-dk64w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.787328 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-xqnrf"] Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.796447 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.874328 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zdqwd" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.883241 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-config\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.883371 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.883670 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnslt\" (UniqueName: \"kubernetes.io/projected/ddd0e60e-a07c-41b0-8c44-00844dfef24a-kube-api-access-vnslt\") pod \"barbican-db-sync-dk64w\" (UID: \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\") " pod="openstack/barbican-db-sync-dk64w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.883703 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.883734 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.883830 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.883852 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnxsb\" (UniqueName: \"kubernetes.io/projected/fce977a9-243a-4707-94ff-2ab3fe9e21d5-kube-api-access-wnxsb\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.883877 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd0e60e-a07c-41b0-8c44-00844dfef24a-combined-ca-bundle\") pod \"barbican-db-sync-dk64w\" (UID: \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\") " pod="openstack/barbican-db-sync-dk64w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.883941 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ddd0e60e-a07c-41b0-8c44-00844dfef24a-db-sync-config-data\") pod \"barbican-db-sync-dk64w\" (UID: \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\") " pod="openstack/barbican-db-sync-dk64w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.893278 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd0e60e-a07c-41b0-8c44-00844dfef24a-combined-ca-bundle\") pod \"barbican-db-sync-dk64w\" (UID: \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\") " pod="openstack/barbican-db-sync-dk64w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.906339 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ddd0e60e-a07c-41b0-8c44-00844dfef24a-db-sync-config-data\") pod \"barbican-db-sync-dk64w\" (UID: \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\") " pod="openstack/barbican-db-sync-dk64w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.914402 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnslt\" (UniqueName: \"kubernetes.io/projected/ddd0e60e-a07c-41b0-8c44-00844dfef24a-kube-api-access-vnslt\") pod \"barbican-db-sync-dk64w\" (UID: \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\") " pod="openstack/barbican-db-sync-dk64w" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.986607 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.986661 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.986722 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.986744 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnxsb\" (UniqueName: \"kubernetes.io/projected/fce977a9-243a-4707-94ff-2ab3fe9e21d5-kube-api-access-wnxsb\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.986924 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-config\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.986960 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.988374 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.989796 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.990576 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.990906 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-config\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:29 crc kubenswrapper[4872]: I0126 09:23:29.993762 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.025828 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnxsb\" (UniqueName: \"kubernetes.io/projected/fce977a9-243a-4707-94ff-2ab3fe9e21d5-kube-api-access-wnxsb\") pod \"dnsmasq-dns-76fcf4b695-xqnrf\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.026307 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dk64w" Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.135944 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.222234 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-m695h"] Jan 26 09:23:30 crc kubenswrapper[4872]: W0126 09:23:30.239381 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1bdce24d_9f71_495b_8995_32365acdb90e.slice/crio-6ea193edd55a76b212b129ed2823d5cdbe2c12ad2d203cec82904b830649031a WatchSource:0}: Error finding container 6ea193edd55a76b212b129ed2823d5cdbe2c12ad2d203cec82904b830649031a: Status 404 returned error can't find the container with id 6ea193edd55a76b212b129ed2823d5cdbe2c12ad2d203cec82904b830649031a Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.240582 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-dhw6k"] Jan 26 09:23:30 crc kubenswrapper[4872]: W0126 09:23:30.242933 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e6f69fe_3bb8_4f5a_bb0c_b0d9a7304688.slice/crio-fa6017553e7cc2bb8db8f23f3afd648b30ccc79c9fe906a8d24f24f8d3e24a94 WatchSource:0}: Error finding container fa6017553e7cc2bb8db8f23f3afd648b30ccc79c9fe906a8d24f24f8d3e24a94: Status 404 returned error can't find the container with id fa6017553e7cc2bb8db8f23f3afd648b30ccc79c9fe906a8d24f24f8d3e24a94 Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.381745 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-create-jdmsk"] Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.482180 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-m695h" event={"ID":"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688","Type":"ContainerStarted","Data":"fa6017553e7cc2bb8db8f23f3afd648b30ccc79c9fe906a8d24f24f8d3e24a94"} Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.484889 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-zd82w"] Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.490462 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dhw6k" event={"ID":"1bdce24d-9f71-495b-8995-32365acdb90e","Type":"ContainerStarted","Data":"6ea193edd55a76b212b129ed2823d5cdbe2c12ad2d203cec82904b830649031a"} Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.495304 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-jdmsk" event={"ID":"de9df0f9-9456-4bfd-98f7-065a933bb80e","Type":"ContainerStarted","Data":"0989438a7afab6ac6ab29e20bf8083efa786aefa4d22b7cac5abb065b65aae42"} Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.496300 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-c3e0-account-create-update-pfn6b"] Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.512637 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-4cbzl"] Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.748584 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zdqwd"] Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.798490 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.853776 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dk64w"] Jan 26 09:23:30 crc kubenswrapper[4872]: I0126 09:23:30.971199 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.001170 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-xqnrf"] Jan 26 09:23:31 crc kubenswrapper[4872]: W0126 09:23:31.004144 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfce977a9_243a_4707_94ff_2ab3fe9e21d5.slice/crio-1a2c7f4b2187eecca1182a3e98533f4dfe2de069e2e1d4adc1b0a49496966f36 WatchSource:0}: Error finding container 1a2c7f4b2187eecca1182a3e98533f4dfe2de069e2e1d4adc1b0a49496966f36: Status 404 returned error can't find the container with id 1a2c7f4b2187eecca1182a3e98533f4dfe2de069e2e1d4adc1b0a49496966f36 Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.508135 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dk64w" event={"ID":"ddd0e60e-a07c-41b0-8c44-00844dfef24a","Type":"ContainerStarted","Data":"92cff79ea1dac3cc7f82afd01745bdedaa7ae94f680e1a2de68b0f4e56c7b2b3"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.510660 4872 generic.go:334] "Generic (PLEG): container finished" podID="fce977a9-243a-4707-94ff-2ab3fe9e21d5" containerID="c0e817426d0cde3c4a3c5418fc2e4e5d41144c9810d7f6309c05ae0e833dbca0" exitCode=0 Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.510945 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" event={"ID":"fce977a9-243a-4707-94ff-2ab3fe9e21d5","Type":"ContainerDied","Data":"c0e817426d0cde3c4a3c5418fc2e4e5d41144c9810d7f6309c05ae0e833dbca0"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.511073 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" event={"ID":"fce977a9-243a-4707-94ff-2ab3fe9e21d5","Type":"ContainerStarted","Data":"1a2c7f4b2187eecca1182a3e98533f4dfe2de069e2e1d4adc1b0a49496966f36"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.521674 4872 generic.go:334] "Generic (PLEG): container finished" podID="abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00" containerID="3eab026336d335767282c539c580990b865721784af21ad1996aa17c6a95c0d4" exitCode=0 Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.521745 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-c3e0-account-create-update-pfn6b" event={"ID":"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00","Type":"ContainerDied","Data":"3eab026336d335767282c539c580990b865721784af21ad1996aa17c6a95c0d4"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.521840 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-c3e0-account-create-update-pfn6b" event={"ID":"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00","Type":"ContainerStarted","Data":"e648b5b12d54c6d629f050064e96e12472852080bb85f1cb23df6cc515e39f4c"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.538937 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dhw6k" event={"ID":"1bdce24d-9f71-495b-8995-32365acdb90e","Type":"ContainerStarted","Data":"d44c7c2ca90b6cfdc141b88d0cec514236df22703ee1b74ced2429b2657294ef"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.542882 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zdqwd" event={"ID":"0a9d0b8b-dad6-4dbd-a261-77264014fc9b","Type":"ContainerStarted","Data":"2b9f2e119c48fdfe5f8713116e7458a787225904d5c9cc664d2d03d7e8c25c52"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.548238 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zd82w" event={"ID":"04076889-57e8-41ad-b47e-6d8e5e23223f","Type":"ContainerStarted","Data":"99a0d1c30dd823ab40c90826b5604682366f97f920ee24739cd9b0346a6f22bd"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.548497 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zd82w" event={"ID":"04076889-57e8-41ad-b47e-6d8e5e23223f","Type":"ContainerStarted","Data":"42a51032703fc00270a2b591fc3de8a2356d6261055f7ceb6919b856ddd8bd19"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.552816 4872 generic.go:334] "Generic (PLEG): container finished" podID="6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688" containerID="6b9f7d5f63b0e44facda0ad88cec296233186b7baf296b1313f870b6edd80746" exitCode=0 Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.553330 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-m695h" event={"ID":"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688","Type":"ContainerDied","Data":"6b9f7d5f63b0e44facda0ad88cec296233186b7baf296b1313f870b6edd80746"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.568394 4872 generic.go:334] "Generic (PLEG): container finished" podID="de9df0f9-9456-4bfd-98f7-065a933bb80e" containerID="6edb450c51c6475eee8f7ffd305e21c37ca6b8ec3c74ce39fd72f7e282bea9b2" exitCode=0 Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.568601 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-jdmsk" event={"ID":"de9df0f9-9456-4bfd-98f7-065a933bb80e","Type":"ContainerDied","Data":"6edb450c51c6475eee8f7ffd305e21c37ca6b8ec3c74ce39fd72f7e282bea9b2"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.584498 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e","Type":"ContainerStarted","Data":"c9e91b0ec85022369e15f303bd12281c20ce037f3a49dd8abf09654957f60857"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.584582 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-dhw6k" podStartSLOduration=3.584567598 podStartE2EDuration="3.584567598s" podCreationTimestamp="2026-01-26 09:23:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:23:31.572059654 +0000 UTC m=+944.880899465" watchObservedRunningTime="2026-01-26 09:23:31.584567598 +0000 UTC m=+944.893407399" Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.590900 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4cbzl" event={"ID":"3f653549-36ce-4db8-9c67-ec95b628e867","Type":"ContainerStarted","Data":"e7a09dbdbfd182e1d55d19f9e397be1330815a9901eb3d38bddf1feaecf09402"} Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.611219 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-zd82w" podStartSLOduration=2.611191478 podStartE2EDuration="2.611191478s" podCreationTimestamp="2026-01-26 09:23:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:23:31.591032211 +0000 UTC m=+944.899872012" watchObservedRunningTime="2026-01-26 09:23:31.611191478 +0000 UTC m=+944.920031269" Jan 26 09:23:31 crc kubenswrapper[4872]: I0126 09:23:31.912866 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.085611 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-459cw\" (UniqueName: \"kubernetes.io/projected/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-kube-api-access-459cw\") pod \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.085683 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-ovsdbserver-nb\") pod \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.085726 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-ovsdbserver-sb\") pod \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.085763 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-dns-svc\") pod \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.087086 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-dns-swift-storage-0\") pod \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.087145 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-config\") pod \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\" (UID: \"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688\") " Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.104441 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-kube-api-access-459cw" (OuterVolumeSpecName: "kube-api-access-459cw") pod "6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688" (UID: "6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688"). InnerVolumeSpecName "kube-api-access-459cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.141011 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688" (UID: "6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.144515 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688" (UID: "6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.145490 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688" (UID: "6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.149933 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688" (UID: "6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.159113 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-config" (OuterVolumeSpecName: "config") pod "6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688" (UID: "6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.189371 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.189418 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.189432 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-459cw\" (UniqueName: \"kubernetes.io/projected/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-kube-api-access-459cw\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.189451 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.189462 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.189474 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.609657 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" event={"ID":"fce977a9-243a-4707-94ff-2ab3fe9e21d5","Type":"ContainerStarted","Data":"d0e4e7512c26a69185eb5e8cfe97260e6071f4182d8f07f0e7403a9ff37251e0"} Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.609791 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.613687 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-m695h" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.613962 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-m695h" event={"ID":"6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688","Type":"ContainerDied","Data":"fa6017553e7cc2bb8db8f23f3afd648b30ccc79c9fe906a8d24f24f8d3e24a94"} Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.614038 4872 scope.go:117] "RemoveContainer" containerID="6b9f7d5f63b0e44facda0ad88cec296233186b7baf296b1313f870b6edd80746" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.617457 4872 generic.go:334] "Generic (PLEG): container finished" podID="342ea714-cff5-4b5b-bf68-f7baf40bd84e" containerID="ad4e458c514b7df57ac246f939d7104cfd24233db841179b21dd81517401bbc0" exitCode=0 Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.617633 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tchz8" event={"ID":"342ea714-cff5-4b5b-bf68-f7baf40bd84e","Type":"ContainerDied","Data":"ad4e458c514b7df57ac246f939d7104cfd24233db841179b21dd81517401bbc0"} Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.642448 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" podStartSLOduration=3.642418859 podStartE2EDuration="3.642418859s" podCreationTimestamp="2026-01-26 09:23:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:23:32.629430123 +0000 UTC m=+945.938269924" watchObservedRunningTime="2026-01-26 09:23:32.642418859 +0000 UTC m=+945.951258660" Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.746125 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-m695h"] Jan 26 09:23:32 crc kubenswrapper[4872]: I0126 09:23:32.757549 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-m695h"] Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.208136 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688" path="/var/lib/kubelet/pods/6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688/volumes" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.263772 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-jdmsk" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.274274 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-c3e0-account-create-update-pfn6b" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.433335 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr6xv\" (UniqueName: \"kubernetes.io/projected/de9df0f9-9456-4bfd-98f7-065a933bb80e-kube-api-access-kr6xv\") pod \"de9df0f9-9456-4bfd-98f7-065a933bb80e\" (UID: \"de9df0f9-9456-4bfd-98f7-065a933bb80e\") " Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.433393 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00-operator-scripts\") pod \"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00\" (UID: \"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00\") " Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.433442 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n4rm\" (UniqueName: \"kubernetes.io/projected/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00-kube-api-access-4n4rm\") pod \"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00\" (UID: \"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00\") " Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.433500 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de9df0f9-9456-4bfd-98f7-065a933bb80e-operator-scripts\") pod \"de9df0f9-9456-4bfd-98f7-065a933bb80e\" (UID: \"de9df0f9-9456-4bfd-98f7-065a933bb80e\") " Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.435096 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de9df0f9-9456-4bfd-98f7-065a933bb80e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "de9df0f9-9456-4bfd-98f7-065a933bb80e" (UID: "de9df0f9-9456-4bfd-98f7-065a933bb80e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.437248 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00" (UID: "abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.446313 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de9df0f9-9456-4bfd-98f7-065a933bb80e-kube-api-access-kr6xv" (OuterVolumeSpecName: "kube-api-access-kr6xv") pod "de9df0f9-9456-4bfd-98f7-065a933bb80e" (UID: "de9df0f9-9456-4bfd-98f7-065a933bb80e"). InnerVolumeSpecName "kube-api-access-kr6xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.460217 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00-kube-api-access-4n4rm" (OuterVolumeSpecName: "kube-api-access-4n4rm") pod "abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00" (UID: "abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00"). InnerVolumeSpecName "kube-api-access-4n4rm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.535552 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr6xv\" (UniqueName: \"kubernetes.io/projected/de9df0f9-9456-4bfd-98f7-065a933bb80e-kube-api-access-kr6xv\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.535592 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.535604 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n4rm\" (UniqueName: \"kubernetes.io/projected/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00-kube-api-access-4n4rm\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.535613 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de9df0f9-9456-4bfd-98f7-065a933bb80e-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.630630 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-c3e0-account-create-update-pfn6b" event={"ID":"abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00","Type":"ContainerDied","Data":"e648b5b12d54c6d629f050064e96e12472852080bb85f1cb23df6cc515e39f4c"} Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.630683 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e648b5b12d54c6d629f050064e96e12472852080bb85f1cb23df6cc515e39f4c" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.630740 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-c3e0-account-create-update-pfn6b" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.648046 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-jdmsk" event={"ID":"de9df0f9-9456-4bfd-98f7-065a933bb80e","Type":"ContainerDied","Data":"0989438a7afab6ac6ab29e20bf8083efa786aefa4d22b7cac5abb065b65aae42"} Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.648093 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-jdmsk" Jan 26 09:23:33 crc kubenswrapper[4872]: I0126 09:23:33.648117 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0989438a7afab6ac6ab29e20bf8083efa786aefa4d22b7cac5abb065b65aae42" Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.132872 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tchz8" Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.256837 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-config-data\") pod \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.257068 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-combined-ca-bundle\") pod \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.257102 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-db-sync-config-data\") pod \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.257188 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfk24\" (UniqueName: \"kubernetes.io/projected/342ea714-cff5-4b5b-bf68-f7baf40bd84e-kube-api-access-kfk24\") pod \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\" (UID: \"342ea714-cff5-4b5b-bf68-f7baf40bd84e\") " Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.265370 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/342ea714-cff5-4b5b-bf68-f7baf40bd84e-kube-api-access-kfk24" (OuterVolumeSpecName: "kube-api-access-kfk24") pod "342ea714-cff5-4b5b-bf68-f7baf40bd84e" (UID: "342ea714-cff5-4b5b-bf68-f7baf40bd84e"). InnerVolumeSpecName "kube-api-access-kfk24". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.265762 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "342ea714-cff5-4b5b-bf68-f7baf40bd84e" (UID: "342ea714-cff5-4b5b-bf68-f7baf40bd84e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.308482 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "342ea714-cff5-4b5b-bf68-f7baf40bd84e" (UID: "342ea714-cff5-4b5b-bf68-f7baf40bd84e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.360248 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.360661 4872 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.360672 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfk24\" (UniqueName: \"kubernetes.io/projected/342ea714-cff5-4b5b-bf68-f7baf40bd84e-kube-api-access-kfk24\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.397236 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-config-data" (OuterVolumeSpecName: "config-data") pod "342ea714-cff5-4b5b-bf68-f7baf40bd84e" (UID: "342ea714-cff5-4b5b-bf68-f7baf40bd84e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.464202 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/342ea714-cff5-4b5b-bf68-f7baf40bd84e-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.666673 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tchz8" event={"ID":"342ea714-cff5-4b5b-bf68-f7baf40bd84e","Type":"ContainerDied","Data":"487783d4c9f452165ab3b1251ca626f255ec01af279284c1418257b2c41ebac9"} Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.666735 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="487783d4c9f452165ab3b1251ca626f255ec01af279284c1418257b2c41ebac9" Jan 26 09:23:34 crc kubenswrapper[4872]: I0126 09:23:34.675825 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tchz8" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.181436 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-xqnrf"] Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.181780 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" podUID="fce977a9-243a-4707-94ff-2ab3fe9e21d5" containerName="dnsmasq-dns" containerID="cri-o://d0e4e7512c26a69185eb5e8cfe97260e6071f4182d8f07f0e7403a9ff37251e0" gracePeriod=10 Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.272081 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-k2frs"] Jan 26 09:23:35 crc kubenswrapper[4872]: E0126 09:23:35.272678 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688" containerName="init" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.272693 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688" containerName="init" Jan 26 09:23:35 crc kubenswrapper[4872]: E0126 09:23:35.272702 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00" containerName="mariadb-account-create-update" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.272709 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00" containerName="mariadb-account-create-update" Jan 26 09:23:35 crc kubenswrapper[4872]: E0126 09:23:35.272731 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9df0f9-9456-4bfd-98f7-065a933bb80e" containerName="mariadb-database-create" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.272738 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9df0f9-9456-4bfd-98f7-065a933bb80e" containerName="mariadb-database-create" Jan 26 09:23:35 crc kubenswrapper[4872]: E0126 09:23:35.272748 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="342ea714-cff5-4b5b-bf68-f7baf40bd84e" containerName="glance-db-sync" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.272753 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="342ea714-cff5-4b5b-bf68-f7baf40bd84e" containerName="glance-db-sync" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.272947 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="342ea714-cff5-4b5b-bf68-f7baf40bd84e" containerName="glance-db-sync" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.272978 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e6f69fe-3bb8-4f5a-bb0c-b0d9a7304688" containerName="init" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.273003 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00" containerName="mariadb-account-create-update" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.273021 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9df0f9-9456-4bfd-98f7-065a933bb80e" containerName="mariadb-database-create" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.274168 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.304844 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-k2frs"] Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.320912 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-config\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.321034 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.321074 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-np59l\" (UniqueName: \"kubernetes.io/projected/eec4c66a-0959-41ef-acaa-44c65d1a9e13-kube-api-access-np59l\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.321143 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.321235 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.321307 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.422338 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-config\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.422709 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.422738 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-np59l\" (UniqueName: \"kubernetes.io/projected/eec4c66a-0959-41ef-acaa-44c65d1a9e13-kube-api-access-np59l\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.422769 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.422791 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.422832 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.423740 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.424373 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-config\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.425044 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.426027 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.426658 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.448995 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-np59l\" (UniqueName: \"kubernetes.io/projected/eec4c66a-0959-41ef-acaa-44c65d1a9e13-kube-api-access-np59l\") pod \"dnsmasq-dns-8b5c85b87-k2frs\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.665255 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.681837 4872 generic.go:334] "Generic (PLEG): container finished" podID="fce977a9-243a-4707-94ff-2ab3fe9e21d5" containerID="d0e4e7512c26a69185eb5e8cfe97260e6071f4182d8f07f0e7403a9ff37251e0" exitCode=0 Jan 26 09:23:35 crc kubenswrapper[4872]: I0126 09:23:35.681894 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" event={"ID":"fce977a9-243a-4707-94ff-2ab3fe9e21d5","Type":"ContainerDied","Data":"d0e4e7512c26a69185eb5e8cfe97260e6071f4182d8f07f0e7403a9ff37251e0"} Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.045399 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.047365 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.050886 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.051067 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.051330 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-n7wxv" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.069539 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.137261 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-scripts\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.137345 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/235e9951-4453-4c1d-b4f1-c62fae1c8e95-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.137385 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/235e9951-4453-4c1d-b4f1-c62fae1c8e95-logs\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.138041 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqk7r\" (UniqueName: \"kubernetes.io/projected/235e9951-4453-4c1d-b4f1-c62fae1c8e95-kube-api-access-lqk7r\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.138374 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-config-data\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.138426 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.138498 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.240483 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.240592 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.240656 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-scripts\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.240721 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/235e9951-4453-4c1d-b4f1-c62fae1c8e95-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.240742 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/235e9951-4453-4c1d-b4f1-c62fae1c8e95-logs\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.240899 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqk7r\" (UniqueName: \"kubernetes.io/projected/235e9951-4453-4c1d-b4f1-c62fae1c8e95-kube-api-access-lqk7r\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.240974 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-config-data\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.241011 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.241402 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/235e9951-4453-4c1d-b4f1-c62fae1c8e95-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.242416 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/235e9951-4453-4c1d-b4f1-c62fae1c8e95-logs\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.245620 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-scripts\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.253167 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.255005 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-config-data\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.262625 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqk7r\" (UniqueName: \"kubernetes.io/projected/235e9951-4453-4c1d-b4f1-c62fae1c8e95-kube-api-access-lqk7r\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.273040 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.384889 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.387172 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.389154 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.395541 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.402649 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.445235 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed6264fe-e625-4a75-bff6-e93e50f581a8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.445716 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.445761 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.445853 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.445882 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed6264fe-e625-4a75-bff6-e93e50f581a8-logs\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.446181 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.446251 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b695f\" (UniqueName: \"kubernetes.io/projected/ed6264fe-e625-4a75-bff6-e93e50f581a8-kube-api-access-b695f\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.547498 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.547549 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.547575 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed6264fe-e625-4a75-bff6-e93e50f581a8-logs\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.547646 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.547670 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b695f\" (UniqueName: \"kubernetes.io/projected/ed6264fe-e625-4a75-bff6-e93e50f581a8-kube-api-access-b695f\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.547709 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed6264fe-e625-4a75-bff6-e93e50f581a8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.547741 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.547993 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.548837 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed6264fe-e625-4a75-bff6-e93e50f581a8-logs\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.548876 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed6264fe-e625-4a75-bff6-e93e50f581a8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.556355 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.583052 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b695f\" (UniqueName: \"kubernetes.io/projected/ed6264fe-e625-4a75-bff6-e93e50f581a8-kube-api-access-b695f\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.592739 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.592939 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.606007 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:36 crc kubenswrapper[4872]: I0126 09:23:36.718545 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 09:23:37 crc kubenswrapper[4872]: I0126 09:23:37.167633 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-k2frs"] Jan 26 09:23:37 crc kubenswrapper[4872]: I0126 09:23:37.180321 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:23:37 crc kubenswrapper[4872]: W0126 09:23:37.188592 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod235e9951_4453_4c1d_b4f1_c62fae1c8e95.slice/crio-4e53b4e9820e759a0f45db6c2952c8d4e1735ef0c30a4122ee31bde307372a7f WatchSource:0}: Error finding container 4e53b4e9820e759a0f45db6c2952c8d4e1735ef0c30a4122ee31bde307372a7f: Status 404 returned error can't find the container with id 4e53b4e9820e759a0f45db6c2952c8d4e1735ef0c30a4122ee31bde307372a7f Jan 26 09:23:37 crc kubenswrapper[4872]: I0126 09:23:37.477658 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:23:37 crc kubenswrapper[4872]: I0126 09:23:37.723270 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"235e9951-4453-4c1d-b4f1-c62fae1c8e95","Type":"ContainerStarted","Data":"4e53b4e9820e759a0f45db6c2952c8d4e1735ef0c30a4122ee31bde307372a7f"} Jan 26 09:23:37 crc kubenswrapper[4872]: I0126 09:23:37.726963 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ed6264fe-e625-4a75-bff6-e93e50f581a8","Type":"ContainerStarted","Data":"c0c45d6aae93305e1a170dfc91b24c439ad86ec719551e059fa69415ae8b756f"} Jan 26 09:23:37 crc kubenswrapper[4872]: I0126 09:23:37.730137 4872 generic.go:334] "Generic (PLEG): container finished" podID="eec4c66a-0959-41ef-acaa-44c65d1a9e13" containerID="fb69ec54392d85cac6272918a6198b93edae1774864b44bf93e91dbdd5783f91" exitCode=0 Jan 26 09:23:37 crc kubenswrapper[4872]: I0126 09:23:37.730172 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" event={"ID":"eec4c66a-0959-41ef-acaa-44c65d1a9e13","Type":"ContainerDied","Data":"fb69ec54392d85cac6272918a6198b93edae1774864b44bf93e91dbdd5783f91"} Jan 26 09:23:37 crc kubenswrapper[4872]: I0126 09:23:37.730188 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" event={"ID":"eec4c66a-0959-41ef-acaa-44c65d1a9e13","Type":"ContainerStarted","Data":"27d5a45559a08b8186990cd3ff92a96521155fcd7c63e66885e2b258668c57fc"} Jan 26 09:23:38 crc kubenswrapper[4872]: I0126 09:23:38.783440 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"235e9951-4453-4c1d-b4f1-c62fae1c8e95","Type":"ContainerStarted","Data":"522ee564370a90f69001a83abfbe9928266d69c61dfe8c34df46d120acdbc217"} Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.126735 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.221005 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.390780 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-db-sync-8dnjd"] Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.402192 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.405373 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-sync-8dnjd"] Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.411821 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-config-data" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.411851 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-dockercfg-6kfkt" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.411821 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-scripts" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.537352 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-combined-ca-bundle\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.537427 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-config-data\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.537490 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3e385f5e-2610-44cd-ac51-ff0642cb47a9-config-data-merged\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.537524 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-scripts\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.537559 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/3e385f5e-2610-44cd-ac51-ff0642cb47a9-etc-podinfo\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.537593 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75gnr\" (UniqueName: \"kubernetes.io/projected/3e385f5e-2610-44cd-ac51-ff0642cb47a9-kube-api-access-75gnr\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.641341 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3e385f5e-2610-44cd-ac51-ff0642cb47a9-config-data-merged\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.641429 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-scripts\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.641481 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/3e385f5e-2610-44cd-ac51-ff0642cb47a9-etc-podinfo\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.641531 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75gnr\" (UniqueName: \"kubernetes.io/projected/3e385f5e-2610-44cd-ac51-ff0642cb47a9-kube-api-access-75gnr\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.641650 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-combined-ca-bundle\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.641700 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-config-data\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.643845 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3e385f5e-2610-44cd-ac51-ff0642cb47a9-config-data-merged\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.650604 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-config-data\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.651292 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-scripts\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.652349 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-combined-ca-bundle\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.661151 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/3e385f5e-2610-44cd-ac51-ff0642cb47a9-etc-podinfo\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.665999 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75gnr\" (UniqueName: \"kubernetes.io/projected/3e385f5e-2610-44cd-ac51-ff0642cb47a9-kube-api-access-75gnr\") pod \"ironic-db-sync-8dnjd\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.730573 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.800135 4872 generic.go:334] "Generic (PLEG): container finished" podID="1bdce24d-9f71-495b-8995-32365acdb90e" containerID="d44c7c2ca90b6cfdc141b88d0cec514236df22703ee1b74ced2429b2657294ef" exitCode=0 Jan 26 09:23:39 crc kubenswrapper[4872]: I0126 09:23:39.800192 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dhw6k" event={"ID":"1bdce24d-9f71-495b-8995-32365acdb90e","Type":"ContainerDied","Data":"d44c7c2ca90b6cfdc141b88d0cec514236df22703ee1b74ced2429b2657294ef"} Jan 26 09:23:40 crc kubenswrapper[4872]: I0126 09:23:40.227438 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:23:40 crc kubenswrapper[4872]: I0126 09:23:40.227870 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:23:44 crc kubenswrapper[4872]: E0126 09:23:44.396865 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Jan 26 09:23:44 crc kubenswrapper[4872]: E0126 09:23:44.398096 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5ltww,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-zdqwd_openstack(0a9d0b8b-dad6-4dbd-a261-77264014fc9b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:23:44 crc kubenswrapper[4872]: E0126 09:23:44.400169 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-zdqwd" podUID="0a9d0b8b-dad6-4dbd-a261-77264014fc9b" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.508278 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.519045 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.670250 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-scripts\") pod \"1bdce24d-9f71-495b-8995-32365acdb90e\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.670337 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-dns-swift-storage-0\") pod \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.670385 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-credential-keys\") pod \"1bdce24d-9f71-495b-8995-32365acdb90e\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.670501 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-combined-ca-bundle\") pod \"1bdce24d-9f71-495b-8995-32365acdb90e\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.670584 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-dns-svc\") pod \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.670627 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-ovsdbserver-nb\") pod \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.670661 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-config\") pod \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.670711 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-ovsdbserver-sb\") pod \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.670789 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mrqg\" (UniqueName: \"kubernetes.io/projected/1bdce24d-9f71-495b-8995-32365acdb90e-kube-api-access-8mrqg\") pod \"1bdce24d-9f71-495b-8995-32365acdb90e\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.670854 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-config-data\") pod \"1bdce24d-9f71-495b-8995-32365acdb90e\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.670887 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-fernet-keys\") pod \"1bdce24d-9f71-495b-8995-32365acdb90e\" (UID: \"1bdce24d-9f71-495b-8995-32365acdb90e\") " Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.670904 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnxsb\" (UniqueName: \"kubernetes.io/projected/fce977a9-243a-4707-94ff-2ab3fe9e21d5-kube-api-access-wnxsb\") pod \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\" (UID: \"fce977a9-243a-4707-94ff-2ab3fe9e21d5\") " Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.678993 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-scripts" (OuterVolumeSpecName: "scripts") pod "1bdce24d-9f71-495b-8995-32365acdb90e" (UID: "1bdce24d-9f71-495b-8995-32365acdb90e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.679711 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1bdce24d-9f71-495b-8995-32365acdb90e" (UID: "1bdce24d-9f71-495b-8995-32365acdb90e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.679818 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bdce24d-9f71-495b-8995-32365acdb90e-kube-api-access-8mrqg" (OuterVolumeSpecName: "kube-api-access-8mrqg") pod "1bdce24d-9f71-495b-8995-32365acdb90e" (UID: "1bdce24d-9f71-495b-8995-32365acdb90e"). InnerVolumeSpecName "kube-api-access-8mrqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.680274 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1bdce24d-9f71-495b-8995-32365acdb90e" (UID: "1bdce24d-9f71-495b-8995-32365acdb90e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.681470 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fce977a9-243a-4707-94ff-2ab3fe9e21d5-kube-api-access-wnxsb" (OuterVolumeSpecName: "kube-api-access-wnxsb") pod "fce977a9-243a-4707-94ff-2ab3fe9e21d5" (UID: "fce977a9-243a-4707-94ff-2ab3fe9e21d5"). InnerVolumeSpecName "kube-api-access-wnxsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.713331 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-config-data" (OuterVolumeSpecName: "config-data") pod "1bdce24d-9f71-495b-8995-32365acdb90e" (UID: "1bdce24d-9f71-495b-8995-32365acdb90e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.736573 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fce977a9-243a-4707-94ff-2ab3fe9e21d5" (UID: "fce977a9-243a-4707-94ff-2ab3fe9e21d5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.736607 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fce977a9-243a-4707-94ff-2ab3fe9e21d5" (UID: "fce977a9-243a-4707-94ff-2ab3fe9e21d5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.745615 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fce977a9-243a-4707-94ff-2ab3fe9e21d5" (UID: "fce977a9-243a-4707-94ff-2ab3fe9e21d5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.745955 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bdce24d-9f71-495b-8995-32365acdb90e" (UID: "1bdce24d-9f71-495b-8995-32365acdb90e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.746960 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fce977a9-243a-4707-94ff-2ab3fe9e21d5" (UID: "fce977a9-243a-4707-94ff-2ab3fe9e21d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.750492 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-config" (OuterVolumeSpecName: "config") pod "fce977a9-243a-4707-94ff-2ab3fe9e21d5" (UID: "fce977a9-243a-4707-94ff-2ab3fe9e21d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.773402 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.773468 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.773483 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.773499 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.773512 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.773523 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mrqg\" (UniqueName: \"kubernetes.io/projected/1bdce24d-9f71-495b-8995-32365acdb90e-kube-api-access-8mrqg\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.773536 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.773543 4872 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.773552 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnxsb\" (UniqueName: \"kubernetes.io/projected/fce977a9-243a-4707-94ff-2ab3fe9e21d5-kube-api-access-wnxsb\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.773562 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.773570 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fce977a9-243a-4707-94ff-2ab3fe9e21d5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.773579 4872 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1bdce24d-9f71-495b-8995-32365acdb90e-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.853710 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" event={"ID":"fce977a9-243a-4707-94ff-2ab3fe9e21d5","Type":"ContainerDied","Data":"1a2c7f4b2187eecca1182a3e98533f4dfe2de069e2e1d4adc1b0a49496966f36"} Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.853729 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.853873 4872 scope.go:117] "RemoveContainer" containerID="d0e4e7512c26a69185eb5e8cfe97260e6071f4182d8f07f0e7403a9ff37251e0" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.856083 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-dhw6k" event={"ID":"1bdce24d-9f71-495b-8995-32365acdb90e","Type":"ContainerDied","Data":"6ea193edd55a76b212b129ed2823d5cdbe2c12ad2d203cec82904b830649031a"} Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.856128 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-dhw6k" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.856142 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ea193edd55a76b212b129ed2823d5cdbe2c12ad2d203cec82904b830649031a" Jan 26 09:23:44 crc kubenswrapper[4872]: E0126 09:23:44.858619 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-zdqwd" podUID="0a9d0b8b-dad6-4dbd-a261-77264014fc9b" Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.911357 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-xqnrf"] Jan 26 09:23:44 crc kubenswrapper[4872]: I0126 09:23:44.919265 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-xqnrf"] Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.138126 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76fcf4b695-xqnrf" podUID="fce977a9-243a-4707-94ff-2ab3fe9e21d5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: i/o timeout" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.198518 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fce977a9-243a-4707-94ff-2ab3fe9e21d5" path="/var/lib/kubelet/pods/fce977a9-243a-4707-94ff-2ab3fe9e21d5/volumes" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.682812 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-dhw6k"] Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.709557 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-dhw6k"] Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.770004 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-p27pg"] Jan 26 09:23:45 crc kubenswrapper[4872]: E0126 09:23:45.770446 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bdce24d-9f71-495b-8995-32365acdb90e" containerName="keystone-bootstrap" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.770472 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bdce24d-9f71-495b-8995-32365acdb90e" containerName="keystone-bootstrap" Jan 26 09:23:45 crc kubenswrapper[4872]: E0126 09:23:45.770489 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fce977a9-243a-4707-94ff-2ab3fe9e21d5" containerName="init" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.770498 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fce977a9-243a-4707-94ff-2ab3fe9e21d5" containerName="init" Jan 26 09:23:45 crc kubenswrapper[4872]: E0126 09:23:45.770516 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fce977a9-243a-4707-94ff-2ab3fe9e21d5" containerName="dnsmasq-dns" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.770525 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fce977a9-243a-4707-94ff-2ab3fe9e21d5" containerName="dnsmasq-dns" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.770742 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fce977a9-243a-4707-94ff-2ab3fe9e21d5" containerName="dnsmasq-dns" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.770758 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bdce24d-9f71-495b-8995-32365acdb90e" containerName="keystone-bootstrap" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.771402 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.776277 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.776495 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.777270 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.777792 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-7k5wq" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.784150 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-p27pg"] Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.909007 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-config-data\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.909064 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f88sc\" (UniqueName: \"kubernetes.io/projected/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-kube-api-access-f88sc\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.909092 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-scripts\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.909477 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-fernet-keys\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.909619 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-combined-ca-bundle\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:45 crc kubenswrapper[4872]: I0126 09:23:45.909682 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-credential-keys\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.011449 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-combined-ca-bundle\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.011521 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-credential-keys\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.011614 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-config-data\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.011642 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f88sc\" (UniqueName: \"kubernetes.io/projected/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-kube-api-access-f88sc\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.011670 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-scripts\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.011763 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-fernet-keys\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.019142 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-fernet-keys\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.019137 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-combined-ca-bundle\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.019653 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-credential-keys\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.020697 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-scripts\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.030134 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-config-data\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.033218 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f88sc\" (UniqueName: \"kubernetes.io/projected/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-kube-api-access-f88sc\") pod \"keystone-bootstrap-p27pg\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:46 crc kubenswrapper[4872]: I0126 09:23:46.097888 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:23:47 crc kubenswrapper[4872]: I0126 09:23:47.204985 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bdce24d-9f71-495b-8995-32365acdb90e" path="/var/lib/kubelet/pods/1bdce24d-9f71-495b-8995-32365acdb90e/volumes" Jan 26 09:23:52 crc kubenswrapper[4872]: E0126 09:23:52.261952 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Jan 26 09:23:52 crc kubenswrapper[4872]: E0126 09:23:52.263840 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vnslt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-dk64w_openstack(ddd0e60e-a07c-41b0-8c44-00844dfef24a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:23:52 crc kubenswrapper[4872]: E0126 09:23:52.265391 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-dk64w" podUID="ddd0e60e-a07c-41b0-8c44-00844dfef24a" Jan 26 09:23:52 crc kubenswrapper[4872]: E0126 09:23:52.941753 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-dk64w" podUID="ddd0e60e-a07c-41b0-8c44-00844dfef24a" Jan 26 09:23:53 crc kubenswrapper[4872]: I0126 09:23:53.956752 4872 scope.go:117] "RemoveContainer" containerID="c0e817426d0cde3c4a3c5418fc2e4e5d41144c9810d7f6309c05ae0e833dbca0" Jan 26 09:23:54 crc kubenswrapper[4872]: E0126 09:23:54.007713 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Jan 26 09:23:54 crc kubenswrapper[4872]: E0126 09:23:54.008025 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kvmw7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-4cbzl_openstack(3f653549-36ce-4db8-9c67-ec95b628e867): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:23:54 crc kubenswrapper[4872]: E0126 09:23:54.009113 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-4cbzl" podUID="3f653549-36ce-4db8-9c67-ec95b628e867" Jan 26 09:23:54 crc kubenswrapper[4872]: I0126 09:23:54.502676 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-sync-8dnjd"] Jan 26 09:23:54 crc kubenswrapper[4872]: I0126 09:23:54.570417 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-p27pg"] Jan 26 09:23:54 crc kubenswrapper[4872]: W0126 09:23:54.577847 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f6ad3ad_8b9d_4dc0_bf7a_7b7eff1e666b.slice/crio-7e83981d05b8ff941a787547a5dfaf491b1761a7f2e533a08f2ac1be52c01c13 WatchSource:0}: Error finding container 7e83981d05b8ff941a787547a5dfaf491b1761a7f2e533a08f2ac1be52c01c13: Status 404 returned error can't find the container with id 7e83981d05b8ff941a787547a5dfaf491b1761a7f2e533a08f2ac1be52c01c13 Jan 26 09:23:54 crc kubenswrapper[4872]: I0126 09:23:54.957038 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-8dnjd" event={"ID":"3e385f5e-2610-44cd-ac51-ff0642cb47a9","Type":"ContainerStarted","Data":"d15a1e6f007a8c6847574b0720848d9f2996d30a3050e89a09a1953bd95ca41a"} Jan 26 09:23:54 crc kubenswrapper[4872]: I0126 09:23:54.960716 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ed6264fe-e625-4a75-bff6-e93e50f581a8","Type":"ContainerStarted","Data":"5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9"} Jan 26 09:23:54 crc kubenswrapper[4872]: I0126 09:23:54.969573 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" event={"ID":"eec4c66a-0959-41ef-acaa-44c65d1a9e13","Type":"ContainerStarted","Data":"2d25c74d651e163d432f82fc0c94cfdf58fdc7cd4cdea162e4e71185f76541ea"} Jan 26 09:23:54 crc kubenswrapper[4872]: I0126 09:23:54.970365 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:23:54 crc kubenswrapper[4872]: I0126 09:23:54.972722 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e","Type":"ContainerStarted","Data":"6aa5198368a8ddf875c04c2c19c0ee31224b4ec54781dc7d825df99106c52642"} Jan 26 09:23:54 crc kubenswrapper[4872]: I0126 09:23:54.975789 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p27pg" event={"ID":"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b","Type":"ContainerStarted","Data":"958517ad3eeeb9238b0d2f145a6ce26ae1a0127d403f1808b713aa7b5df84638"} Jan 26 09:23:54 crc kubenswrapper[4872]: I0126 09:23:54.975885 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p27pg" event={"ID":"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b","Type":"ContainerStarted","Data":"7e83981d05b8ff941a787547a5dfaf491b1761a7f2e533a08f2ac1be52c01c13"} Jan 26 09:23:54 crc kubenswrapper[4872]: I0126 09:23:54.979932 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="235e9951-4453-4c1d-b4f1-c62fae1c8e95" containerName="glance-log" containerID="cri-o://522ee564370a90f69001a83abfbe9928266d69c61dfe8c34df46d120acdbc217" gracePeriod=30 Jan 26 09:23:54 crc kubenswrapper[4872]: I0126 09:23:54.980002 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="235e9951-4453-4c1d-b4f1-c62fae1c8e95" containerName="glance-httpd" containerID="cri-o://dc9c14428a00367e9f835b6a4447baa4d962f072165f29de31eca98959b809c3" gracePeriod=30 Jan 26 09:23:54 crc kubenswrapper[4872]: I0126 09:23:54.979926 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"235e9951-4453-4c1d-b4f1-c62fae1c8e95","Type":"ContainerStarted","Data":"dc9c14428a00367e9f835b6a4447baa4d962f072165f29de31eca98959b809c3"} Jan 26 09:23:54 crc kubenswrapper[4872]: E0126 09:23:54.982313 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-4cbzl" podUID="3f653549-36ce-4db8-9c67-ec95b628e867" Jan 26 09:23:55 crc kubenswrapper[4872]: I0126 09:23:55.005550 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" podStartSLOduration=20.005519164 podStartE2EDuration="20.005519164s" podCreationTimestamp="2026-01-26 09:23:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:23:54.996757399 +0000 UTC m=+968.305597200" watchObservedRunningTime="2026-01-26 09:23:55.005519164 +0000 UTC m=+968.314358965" Jan 26 09:23:55 crc kubenswrapper[4872]: I0126 09:23:55.026252 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-p27pg" podStartSLOduration=10.026226985 podStartE2EDuration="10.026226985s" podCreationTimestamp="2026-01-26 09:23:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:23:55.01469663 +0000 UTC m=+968.323536431" watchObservedRunningTime="2026-01-26 09:23:55.026226985 +0000 UTC m=+968.335066786" Jan 26 09:23:55 crc kubenswrapper[4872]: I0126 09:23:55.063776 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=20.063741179 podStartE2EDuration="20.063741179s" podCreationTimestamp="2026-01-26 09:23:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:23:55.057914658 +0000 UTC m=+968.366754459" watchObservedRunningTime="2026-01-26 09:23:55.063741179 +0000 UTC m=+968.372580980" Jan 26 09:23:55 crc kubenswrapper[4872]: I0126 09:23:55.994439 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ed6264fe-e625-4a75-bff6-e93e50f581a8","Type":"ContainerStarted","Data":"cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835"} Jan 26 09:23:55 crc kubenswrapper[4872]: I0126 09:23:55.994589 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ed6264fe-e625-4a75-bff6-e93e50f581a8" containerName="glance-log" containerID="cri-o://5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9" gracePeriod=30 Jan 26 09:23:55 crc kubenswrapper[4872]: I0126 09:23:55.994713 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ed6264fe-e625-4a75-bff6-e93e50f581a8" containerName="glance-httpd" containerID="cri-o://cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835" gracePeriod=30 Jan 26 09:23:55 crc kubenswrapper[4872]: I0126 09:23:55.997847 4872 generic.go:334] "Generic (PLEG): container finished" podID="235e9951-4453-4c1d-b4f1-c62fae1c8e95" containerID="dc9c14428a00367e9f835b6a4447baa4d962f072165f29de31eca98959b809c3" exitCode=0 Jan 26 09:23:55 crc kubenswrapper[4872]: I0126 09:23:55.997892 4872 generic.go:334] "Generic (PLEG): container finished" podID="235e9951-4453-4c1d-b4f1-c62fae1c8e95" containerID="522ee564370a90f69001a83abfbe9928266d69c61dfe8c34df46d120acdbc217" exitCode=143 Jan 26 09:23:55 crc kubenswrapper[4872]: I0126 09:23:55.997943 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"235e9951-4453-4c1d-b4f1-c62fae1c8e95","Type":"ContainerDied","Data":"dc9c14428a00367e9f835b6a4447baa4d962f072165f29de31eca98959b809c3"} Jan 26 09:23:55 crc kubenswrapper[4872]: I0126 09:23:55.997992 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"235e9951-4453-4c1d-b4f1-c62fae1c8e95","Type":"ContainerDied","Data":"522ee564370a90f69001a83abfbe9928266d69c61dfe8c34df46d120acdbc217"} Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.023456 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=21.023431179 podStartE2EDuration="21.023431179s" podCreationTimestamp="2026-01-26 09:23:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:23:56.022771042 +0000 UTC m=+969.331610843" watchObservedRunningTime="2026-01-26 09:23:56.023431179 +0000 UTC m=+969.332270980" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.283849 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.344388 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-config-data\") pod \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.344488 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqk7r\" (UniqueName: \"kubernetes.io/projected/235e9951-4453-4c1d-b4f1-c62fae1c8e95-kube-api-access-lqk7r\") pod \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.344605 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/235e9951-4453-4c1d-b4f1-c62fae1c8e95-httpd-run\") pod \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.344751 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-combined-ca-bundle\") pod \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.344914 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/235e9951-4453-4c1d-b4f1-c62fae1c8e95-logs\") pod \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.344995 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.345065 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-scripts\") pod \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\" (UID: \"235e9951-4453-4c1d-b4f1-c62fae1c8e95\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.345930 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/235e9951-4453-4c1d-b4f1-c62fae1c8e95-logs" (OuterVolumeSpecName: "logs") pod "235e9951-4453-4c1d-b4f1-c62fae1c8e95" (UID: "235e9951-4453-4c1d-b4f1-c62fae1c8e95"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.346460 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/235e9951-4453-4c1d-b4f1-c62fae1c8e95-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "235e9951-4453-4c1d-b4f1-c62fae1c8e95" (UID: "235e9951-4453-4c1d-b4f1-c62fae1c8e95"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.352367 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/235e9951-4453-4c1d-b4f1-c62fae1c8e95-kube-api-access-lqk7r" (OuterVolumeSpecName: "kube-api-access-lqk7r") pod "235e9951-4453-4c1d-b4f1-c62fae1c8e95" (UID: "235e9951-4453-4c1d-b4f1-c62fae1c8e95"). InnerVolumeSpecName "kube-api-access-lqk7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.352849 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "235e9951-4453-4c1d-b4f1-c62fae1c8e95" (UID: "235e9951-4453-4c1d-b4f1-c62fae1c8e95"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.362273 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-scripts" (OuterVolumeSpecName: "scripts") pod "235e9951-4453-4c1d-b4f1-c62fae1c8e95" (UID: "235e9951-4453-4c1d-b4f1-c62fae1c8e95"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.410206 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "235e9951-4453-4c1d-b4f1-c62fae1c8e95" (UID: "235e9951-4453-4c1d-b4f1-c62fae1c8e95"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.442973 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-config-data" (OuterVolumeSpecName: "config-data") pod "235e9951-4453-4c1d-b4f1-c62fae1c8e95" (UID: "235e9951-4453-4c1d-b4f1-c62fae1c8e95"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.448569 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqk7r\" (UniqueName: \"kubernetes.io/projected/235e9951-4453-4c1d-b4f1-c62fae1c8e95-kube-api-access-lqk7r\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.448612 4872 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/235e9951-4453-4c1d-b4f1-c62fae1c8e95-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.448624 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.448634 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/235e9951-4453-4c1d-b4f1-c62fae1c8e95-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.448679 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.448689 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.448698 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/235e9951-4453-4c1d-b4f1-c62fae1c8e95-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.479326 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.550002 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.673182 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.753462 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed6264fe-e625-4a75-bff6-e93e50f581a8-logs\") pod \"ed6264fe-e625-4a75-bff6-e93e50f581a8\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.753651 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-scripts\") pod \"ed6264fe-e625-4a75-bff6-e93e50f581a8\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.753711 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed6264fe-e625-4a75-bff6-e93e50f581a8-httpd-run\") pod \"ed6264fe-e625-4a75-bff6-e93e50f581a8\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.753762 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ed6264fe-e625-4a75-bff6-e93e50f581a8\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.753827 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-combined-ca-bundle\") pod \"ed6264fe-e625-4a75-bff6-e93e50f581a8\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.753916 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b695f\" (UniqueName: \"kubernetes.io/projected/ed6264fe-e625-4a75-bff6-e93e50f581a8-kube-api-access-b695f\") pod \"ed6264fe-e625-4a75-bff6-e93e50f581a8\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.753949 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-config-data\") pod \"ed6264fe-e625-4a75-bff6-e93e50f581a8\" (UID: \"ed6264fe-e625-4a75-bff6-e93e50f581a8\") " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.754412 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed6264fe-e625-4a75-bff6-e93e50f581a8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ed6264fe-e625-4a75-bff6-e93e50f581a8" (UID: "ed6264fe-e625-4a75-bff6-e93e50f581a8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.754444 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed6264fe-e625-4a75-bff6-e93e50f581a8-logs" (OuterVolumeSpecName: "logs") pod "ed6264fe-e625-4a75-bff6-e93e50f581a8" (UID: "ed6264fe-e625-4a75-bff6-e93e50f581a8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.760038 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-scripts" (OuterVolumeSpecName: "scripts") pod "ed6264fe-e625-4a75-bff6-e93e50f581a8" (UID: "ed6264fe-e625-4a75-bff6-e93e50f581a8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.760510 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "ed6264fe-e625-4a75-bff6-e93e50f581a8" (UID: "ed6264fe-e625-4a75-bff6-e93e50f581a8"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.761131 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed6264fe-e625-4a75-bff6-e93e50f581a8-kube-api-access-b695f" (OuterVolumeSpecName: "kube-api-access-b695f") pod "ed6264fe-e625-4a75-bff6-e93e50f581a8" (UID: "ed6264fe-e625-4a75-bff6-e93e50f581a8"). InnerVolumeSpecName "kube-api-access-b695f". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.782075 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed6264fe-e625-4a75-bff6-e93e50f581a8" (UID: "ed6264fe-e625-4a75-bff6-e93e50f581a8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.803242 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-config-data" (OuterVolumeSpecName: "config-data") pod "ed6264fe-e625-4a75-bff6-e93e50f581a8" (UID: "ed6264fe-e625-4a75-bff6-e93e50f581a8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.855574 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.855603 4872 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ed6264fe-e625-4a75-bff6-e93e50f581a8-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.855638 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.855649 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.855661 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b695f\" (UniqueName: \"kubernetes.io/projected/ed6264fe-e625-4a75-bff6-e93e50f581a8-kube-api-access-b695f\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.855670 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed6264fe-e625-4a75-bff6-e93e50f581a8-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.855679 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed6264fe-e625-4a75-bff6-e93e50f581a8-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.873438 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Jan 26 09:23:56 crc kubenswrapper[4872]: I0126 09:23:56.957215 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.016104 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zdqwd" event={"ID":"0a9d0b8b-dad6-4dbd-a261-77264014fc9b","Type":"ContainerStarted","Data":"14c4e3f971586e364f1d40ef94a38c377b3983ed0b4ecf9f6125b69448b454d4"} Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.032289 4872 generic.go:334] "Generic (PLEG): container finished" podID="ed6264fe-e625-4a75-bff6-e93e50f581a8" containerID="cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835" exitCode=0 Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.032375 4872 generic.go:334] "Generic (PLEG): container finished" podID="ed6264fe-e625-4a75-bff6-e93e50f581a8" containerID="5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9" exitCode=143 Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.032546 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ed6264fe-e625-4a75-bff6-e93e50f581a8","Type":"ContainerDied","Data":"cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835"} Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.032631 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ed6264fe-e625-4a75-bff6-e93e50f581a8","Type":"ContainerDied","Data":"5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9"} Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.032660 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ed6264fe-e625-4a75-bff6-e93e50f581a8","Type":"ContainerDied","Data":"c0c45d6aae93305e1a170dfc91b24c439ad86ec719551e059fa69415ae8b756f"} Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.032720 4872 scope.go:117] "RemoveContainer" containerID="cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.035474 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.057943 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e","Type":"ContainerStarted","Data":"02b520f81790fe052d10b40bbca3e30c9b45bb327c1aab132b4362a8bbda7d52"} Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.066637 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-zdqwd" podStartSLOduration=2.076583499 podStartE2EDuration="28.066614652s" podCreationTimestamp="2026-01-26 09:23:29 +0000 UTC" firstStartedPulling="2026-01-26 09:23:30.784294505 +0000 UTC m=+944.093134306" lastFinishedPulling="2026-01-26 09:23:56.774325658 +0000 UTC m=+970.083165459" observedRunningTime="2026-01-26 09:23:57.056977235 +0000 UTC m=+970.365817056" watchObservedRunningTime="2026-01-26 09:23:57.066614652 +0000 UTC m=+970.375454453" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.068792 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"235e9951-4453-4c1d-b4f1-c62fae1c8e95","Type":"ContainerDied","Data":"4e53b4e9820e759a0f45db6c2952c8d4e1735ef0c30a4122ee31bde307372a7f"} Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.068943 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.171688 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.202239 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.226037 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.235406 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.241095 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:23:57 crc kubenswrapper[4872]: E0126 09:23:57.241596 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed6264fe-e625-4a75-bff6-e93e50f581a8" containerName="glance-log" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.241630 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed6264fe-e625-4a75-bff6-e93e50f581a8" containerName="glance-log" Jan 26 09:23:57 crc kubenswrapper[4872]: E0126 09:23:57.241653 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235e9951-4453-4c1d-b4f1-c62fae1c8e95" containerName="glance-httpd" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.241663 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="235e9951-4453-4c1d-b4f1-c62fae1c8e95" containerName="glance-httpd" Jan 26 09:23:57 crc kubenswrapper[4872]: E0126 09:23:57.241689 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed6264fe-e625-4a75-bff6-e93e50f581a8" containerName="glance-httpd" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.241695 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed6264fe-e625-4a75-bff6-e93e50f581a8" containerName="glance-httpd" Jan 26 09:23:57 crc kubenswrapper[4872]: E0126 09:23:57.241707 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235e9951-4453-4c1d-b4f1-c62fae1c8e95" containerName="glance-log" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.241714 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="235e9951-4453-4c1d-b4f1-c62fae1c8e95" containerName="glance-log" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.241917 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="235e9951-4453-4c1d-b4f1-c62fae1c8e95" containerName="glance-httpd" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.241941 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="235e9951-4453-4c1d-b4f1-c62fae1c8e95" containerName="glance-log" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.241955 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed6264fe-e625-4a75-bff6-e93e50f581a8" containerName="glance-httpd" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.241967 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed6264fe-e625-4a75-bff6-e93e50f581a8" containerName="glance-log" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.244096 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.252551 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.252774 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-n7wxv" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.253571 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.255110 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.275562 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.277407 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.281793 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.281953 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.295941 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.321599 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.368158 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.368304 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.368333 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.368353 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.368380 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.368663 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.368763 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7856628-f316-416e-a721-bb74270c6a02-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.368828 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtwq5\" (UniqueName: \"kubernetes.io/projected/a7856628-f316-416e-a721-bb74270c6a02-kube-api-access-vtwq5\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.368861 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.368963 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-config-data\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.369009 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d24cd308-eaee-4bf0-be25-f4822e095b22-logs\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.369191 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d24cd308-eaee-4bf0-be25-f4822e095b22-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.369339 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx9m5\" (UniqueName: \"kubernetes.io/projected/d24cd308-eaee-4bf0-be25-f4822e095b22-kube-api-access-cx9m5\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.369380 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-scripts\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.369441 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.369518 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7856628-f316-416e-a721-bb74270c6a02-logs\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471451 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471544 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471568 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7856628-f316-416e-a721-bb74270c6a02-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471594 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471616 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtwq5\" (UniqueName: \"kubernetes.io/projected/a7856628-f316-416e-a721-bb74270c6a02-kube-api-access-vtwq5\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471668 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-config-data\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471689 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d24cd308-eaee-4bf0-be25-f4822e095b22-logs\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471743 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d24cd308-eaee-4bf0-be25-f4822e095b22-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471783 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx9m5\" (UniqueName: \"kubernetes.io/projected/d24cd308-eaee-4bf0-be25-f4822e095b22-kube-api-access-cx9m5\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471825 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-scripts\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471851 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471885 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7856628-f316-416e-a721-bb74270c6a02-logs\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471920 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.471975 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.472002 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.472026 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.473342 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d24cd308-eaee-4bf0-be25-f4822e095b22-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.473602 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.474467 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7856628-f316-416e-a721-bb74270c6a02-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.474491 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7856628-f316-416e-a721-bb74270c6a02-logs\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.475477 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.479331 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.480978 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d24cd308-eaee-4bf0-be25-f4822e095b22-logs\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.486096 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.490160 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-scripts\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.492675 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.492973 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.493697 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.500225 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx9m5\" (UniqueName: \"kubernetes.io/projected/d24cd308-eaee-4bf0-be25-f4822e095b22-kube-api-access-cx9m5\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.500724 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtwq5\" (UniqueName: \"kubernetes.io/projected/a7856628-f316-416e-a721-bb74270c6a02-kube-api-access-vtwq5\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.510010 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.538452 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.539018 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-config-data\") pod \"glance-default-external-api-0\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.561700 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.584107 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 09:23:57 crc kubenswrapper[4872]: I0126 09:23:57.612701 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 09:23:58 crc kubenswrapper[4872]: I0126 09:23:58.082242 4872 generic.go:334] "Generic (PLEG): container finished" podID="04076889-57e8-41ad-b47e-6d8e5e23223f" containerID="99a0d1c30dd823ab40c90826b5604682366f97f920ee24739cd9b0346a6f22bd" exitCode=0 Jan 26 09:23:58 crc kubenswrapper[4872]: I0126 09:23:58.082748 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zd82w" event={"ID":"04076889-57e8-41ad-b47e-6d8e5e23223f","Type":"ContainerDied","Data":"99a0d1c30dd823ab40c90826b5604682366f97f920ee24739cd9b0346a6f22bd"} Jan 26 09:23:59 crc kubenswrapper[4872]: I0126 09:23:59.099372 4872 generic.go:334] "Generic (PLEG): container finished" podID="0a9d0b8b-dad6-4dbd-a261-77264014fc9b" containerID="14c4e3f971586e364f1d40ef94a38c377b3983ed0b4ecf9f6125b69448b454d4" exitCode=0 Jan 26 09:23:59 crc kubenswrapper[4872]: I0126 09:23:59.099482 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zdqwd" event={"ID":"0a9d0b8b-dad6-4dbd-a261-77264014fc9b","Type":"ContainerDied","Data":"14c4e3f971586e364f1d40ef94a38c377b3983ed0b4ecf9f6125b69448b454d4"} Jan 26 09:23:59 crc kubenswrapper[4872]: I0126 09:23:59.105547 4872 generic.go:334] "Generic (PLEG): container finished" podID="1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b" containerID="958517ad3eeeb9238b0d2f145a6ce26ae1a0127d403f1808b713aa7b5df84638" exitCode=0 Jan 26 09:23:59 crc kubenswrapper[4872]: I0126 09:23:59.105914 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p27pg" event={"ID":"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b","Type":"ContainerDied","Data":"958517ad3eeeb9238b0d2f145a6ce26ae1a0127d403f1808b713aa7b5df84638"} Jan 26 09:23:59 crc kubenswrapper[4872]: I0126 09:23:59.200200 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="235e9951-4453-4c1d-b4f1-c62fae1c8e95" path="/var/lib/kubelet/pods/235e9951-4453-4c1d-b4f1-c62fae1c8e95/volumes" Jan 26 09:23:59 crc kubenswrapper[4872]: I0126 09:23:59.201211 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed6264fe-e625-4a75-bff6-e93e50f581a8" path="/var/lib/kubelet/pods/ed6264fe-e625-4a75-bff6-e93e50f581a8/volumes" Jan 26 09:24:00 crc kubenswrapper[4872]: I0126 09:24:00.668044 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:24:00 crc kubenswrapper[4872]: I0126 09:24:00.731355 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-p9gpz"] Jan 26 09:24:00 crc kubenswrapper[4872]: I0126 09:24:00.731671 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" podUID="1b1367db-cf5f-45c6-b0e6-b9dee6095824" containerName="dnsmasq-dns" containerID="cri-o://fd481a6bd06ffca5a8b856db10d95109dd0a586ba26be602f072a9d9c42c1a10" gracePeriod=10 Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.133513 4872 generic.go:334] "Generic (PLEG): container finished" podID="1b1367db-cf5f-45c6-b0e6-b9dee6095824" containerID="fd481a6bd06ffca5a8b856db10d95109dd0a586ba26be602f072a9d9c42c1a10" exitCode=0 Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.133702 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" event={"ID":"1b1367db-cf5f-45c6-b0e6-b9dee6095824","Type":"ContainerDied","Data":"fd481a6bd06ffca5a8b856db10d95109dd0a586ba26be602f072a9d9c42c1a10"} Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.243253 4872 scope.go:117] "RemoveContainer" containerID="5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.281873 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zdqwd" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.283886 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zd82w" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.372424 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-logs\") pod \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.373023 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-logs" (OuterVolumeSpecName: "logs") pod "0a9d0b8b-dad6-4dbd-a261-77264014fc9b" (UID: "0a9d0b8b-dad6-4dbd-a261-77264014fc9b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.373197 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04076889-57e8-41ad-b47e-6d8e5e23223f-combined-ca-bundle\") pod \"04076889-57e8-41ad-b47e-6d8e5e23223f\" (UID: \"04076889-57e8-41ad-b47e-6d8e5e23223f\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.373344 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-scripts\") pod \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.373385 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/04076889-57e8-41ad-b47e-6d8e5e23223f-config\") pod \"04076889-57e8-41ad-b47e-6d8e5e23223f\" (UID: \"04076889-57e8-41ad-b47e-6d8e5e23223f\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.373433 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8tv7\" (UniqueName: \"kubernetes.io/projected/04076889-57e8-41ad-b47e-6d8e5e23223f-kube-api-access-v8tv7\") pod \"04076889-57e8-41ad-b47e-6d8e5e23223f\" (UID: \"04076889-57e8-41ad-b47e-6d8e5e23223f\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.373596 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-combined-ca-bundle\") pod \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.373638 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-config-data\") pod \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.373670 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ltww\" (UniqueName: \"kubernetes.io/projected/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-kube-api-access-5ltww\") pod \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\" (UID: \"0a9d0b8b-dad6-4dbd-a261-77264014fc9b\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.374138 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.382050 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-scripts" (OuterVolumeSpecName: "scripts") pod "0a9d0b8b-dad6-4dbd-a261-77264014fc9b" (UID: "0a9d0b8b-dad6-4dbd-a261-77264014fc9b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.384210 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04076889-57e8-41ad-b47e-6d8e5e23223f-kube-api-access-v8tv7" (OuterVolumeSpecName: "kube-api-access-v8tv7") pod "04076889-57e8-41ad-b47e-6d8e5e23223f" (UID: "04076889-57e8-41ad-b47e-6d8e5e23223f"). InnerVolumeSpecName "kube-api-access-v8tv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.393984 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.398151 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-kube-api-access-5ltww" (OuterVolumeSpecName: "kube-api-access-5ltww") pod "0a9d0b8b-dad6-4dbd-a261-77264014fc9b" (UID: "0a9d0b8b-dad6-4dbd-a261-77264014fc9b"). InnerVolumeSpecName "kube-api-access-5ltww". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.412778 4872 scope.go:117] "RemoveContainer" containerID="cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835" Jan 26 09:24:01 crc kubenswrapper[4872]: E0126 09:24:01.413929 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835\": container with ID starting with cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835 not found: ID does not exist" containerID="cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.413994 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835"} err="failed to get container status \"cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835\": rpc error: code = NotFound desc = could not find container \"cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835\": container with ID starting with cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835 not found: ID does not exist" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.414023 4872 scope.go:117] "RemoveContainer" containerID="5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9" Jan 26 09:24:01 crc kubenswrapper[4872]: E0126 09:24:01.415388 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9\": container with ID starting with 5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9 not found: ID does not exist" containerID="5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.415449 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9"} err="failed to get container status \"5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9\": rpc error: code = NotFound desc = could not find container \"5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9\": container with ID starting with 5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9 not found: ID does not exist" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.415468 4872 scope.go:117] "RemoveContainer" containerID="cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.416492 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835"} err="failed to get container status \"cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835\": rpc error: code = NotFound desc = could not find container \"cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835\": container with ID starting with cdeed6e8f5b9173d8321657232aea7f01a90493685d2f609f0c09fb794114835 not found: ID does not exist" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.416519 4872 scope.go:117] "RemoveContainer" containerID="5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.417199 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9"} err="failed to get container status \"5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9\": rpc error: code = NotFound desc = could not find container \"5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9\": container with ID starting with 5f307fdc1a8dcab7fc1241df2da7faad0dd831299732b5b21c4a53d34b27aac9 not found: ID does not exist" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.417255 4872 scope.go:117] "RemoveContainer" containerID="dc9c14428a00367e9f835b6a4447baa4d962f072165f29de31eca98959b809c3" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.426937 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04076889-57e8-41ad-b47e-6d8e5e23223f-config" (OuterVolumeSpecName: "config") pod "04076889-57e8-41ad-b47e-6d8e5e23223f" (UID: "04076889-57e8-41ad-b47e-6d8e5e23223f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.468769 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a9d0b8b-dad6-4dbd-a261-77264014fc9b" (UID: "0a9d0b8b-dad6-4dbd-a261-77264014fc9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.494705 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04076889-57e8-41ad-b47e-6d8e5e23223f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04076889-57e8-41ad-b47e-6d8e5e23223f" (UID: "04076889-57e8-41ad-b47e-6d8e5e23223f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.498724 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-config-data" (OuterVolumeSpecName: "config-data") pod "0a9d0b8b-dad6-4dbd-a261-77264014fc9b" (UID: "0a9d0b8b-dad6-4dbd-a261-77264014fc9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.501930 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-combined-ca-bundle\") pod \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.502079 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-credential-keys\") pod \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.502169 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f88sc\" (UniqueName: \"kubernetes.io/projected/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-kube-api-access-f88sc\") pod \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.502250 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-config-data\") pod \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.502357 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-fernet-keys\") pod \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.502390 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-scripts\") pod \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\" (UID: \"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.509988 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.510034 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/04076889-57e8-41ad-b47e-6d8e5e23223f-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.510046 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8tv7\" (UniqueName: \"kubernetes.io/projected/04076889-57e8-41ad-b47e-6d8e5e23223f-kube-api-access-v8tv7\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.510062 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.510078 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.510090 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ltww\" (UniqueName: \"kubernetes.io/projected/0a9d0b8b-dad6-4dbd-a261-77264014fc9b-kube-api-access-5ltww\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.510100 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04076889-57e8-41ad-b47e-6d8e5e23223f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.515390 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b" (UID: "1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.518907 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.550090 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-scripts" (OuterVolumeSpecName: "scripts") pod "1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b" (UID: "1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.555594 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-kube-api-access-f88sc" (OuterVolumeSpecName: "kube-api-access-f88sc") pod "1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b" (UID: "1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b"). InnerVolumeSpecName "kube-api-access-f88sc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.558317 4872 scope.go:117] "RemoveContainer" containerID="522ee564370a90f69001a83abfbe9928266d69c61dfe8c34df46d120acdbc217" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.559197 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b" (UID: "1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.564470 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-config-data" (OuterVolumeSpecName: "config-data") pod "1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b" (UID: "1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.575302 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b" (UID: "1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.610836 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-dns-svc\") pod \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.610891 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-ovsdbserver-nb\") pod \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.610986 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-ovsdbserver-sb\") pod \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.611030 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dvjf\" (UniqueName: \"kubernetes.io/projected/1b1367db-cf5f-45c6-b0e6-b9dee6095824-kube-api-access-7dvjf\") pod \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.611050 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-config\") pod \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.611148 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-dns-swift-storage-0\") pod \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\" (UID: \"1b1367db-cf5f-45c6-b0e6-b9dee6095824\") " Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.611486 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.611498 4872 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-credential-keys\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.611507 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f88sc\" (UniqueName: \"kubernetes.io/projected/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-kube-api-access-f88sc\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.611519 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.611528 4872 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-fernet-keys\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.611536 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.623592 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b1367db-cf5f-45c6-b0e6-b9dee6095824-kube-api-access-7dvjf" (OuterVolumeSpecName: "kube-api-access-7dvjf") pod "1b1367db-cf5f-45c6-b0e6-b9dee6095824" (UID: "1b1367db-cf5f-45c6-b0e6-b9dee6095824"). InnerVolumeSpecName "kube-api-access-7dvjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.670480 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b1367db-cf5f-45c6-b0e6-b9dee6095824" (UID: "1b1367db-cf5f-45c6-b0e6-b9dee6095824"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.676005 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-config" (OuterVolumeSpecName: "config") pod "1b1367db-cf5f-45c6-b0e6-b9dee6095824" (UID: "1b1367db-cf5f-45c6-b0e6-b9dee6095824"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.685052 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1b1367db-cf5f-45c6-b0e6-b9dee6095824" (UID: "1b1367db-cf5f-45c6-b0e6-b9dee6095824"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.705752 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1b1367db-cf5f-45c6-b0e6-b9dee6095824" (UID: "1b1367db-cf5f-45c6-b0e6-b9dee6095824"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.713408 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.713690 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.714157 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.714344 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dvjf\" (UniqueName: \"kubernetes.io/projected/1b1367db-cf5f-45c6-b0e6-b9dee6095824-kube-api-access-7dvjf\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.714476 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.715850 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1b1367db-cf5f-45c6-b0e6-b9dee6095824" (UID: "1b1367db-cf5f-45c6-b0e6-b9dee6095824"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.752525 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:24:01 crc kubenswrapper[4872]: I0126 09:24:01.816710 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b1367db-cf5f-45c6-b0e6-b9dee6095824-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.005469 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:24:02 crc kubenswrapper[4872]: W0126 09:24:02.020940 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd24cd308_eaee_4bf0_be25_f4822e095b22.slice/crio-8e29cbee17b05120d742926e8135c716093985cd80e6b003d187a062ec5475c2 WatchSource:0}: Error finding container 8e29cbee17b05120d742926e8135c716093985cd80e6b003d187a062ec5475c2: Status 404 returned error can't find the container with id 8e29cbee17b05120d742926e8135c716093985cd80e6b003d187a062ec5475c2 Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.171937 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zdqwd" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.171937 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zdqwd" event={"ID":"0a9d0b8b-dad6-4dbd-a261-77264014fc9b","Type":"ContainerDied","Data":"2b9f2e119c48fdfe5f8713116e7458a787225904d5c9cc664d2d03d7e8c25c52"} Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.172112 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b9f2e119c48fdfe5f8713116e7458a787225904d5c9cc664d2d03d7e8c25c52" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.179640 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" event={"ID":"1b1367db-cf5f-45c6-b0e6-b9dee6095824","Type":"ContainerDied","Data":"def2322ba18e2adaadc8ab0504e828186e6bfe0b3fc9667587a1a712698127c5"} Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.179717 4872 scope.go:117] "RemoveContainer" containerID="fd481a6bd06ffca5a8b856db10d95109dd0a586ba26be602f072a9d9c42c1a10" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.179752 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-p9gpz" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.187660 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p27pg" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.187669 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p27pg" event={"ID":"1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b","Type":"ContainerDied","Data":"7e83981d05b8ff941a787547a5dfaf491b1761a7f2e533a08f2ac1be52c01c13"} Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.188282 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e83981d05b8ff941a787547a5dfaf491b1761a7f2e533a08f2ac1be52c01c13" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.191092 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a7856628-f316-416e-a721-bb74270c6a02","Type":"ContainerStarted","Data":"5f94a99ecf19d7abf57d2cd9f49d9d1bfcf1492bbf84aaebbc7e0ea7789d475a"} Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.194474 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d24cd308-eaee-4bf0-be25-f4822e095b22","Type":"ContainerStarted","Data":"8e29cbee17b05120d742926e8135c716093985cd80e6b003d187a062ec5475c2"} Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.196668 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-zd82w" event={"ID":"04076889-57e8-41ad-b47e-6d8e5e23223f","Type":"ContainerDied","Data":"42a51032703fc00270a2b591fc3de8a2356d6261055f7ceb6919b856ddd8bd19"} Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.196699 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42a51032703fc00270a2b591fc3de8a2356d6261055f7ceb6919b856ddd8bd19" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.196782 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-zd82w" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.201165 4872 generic.go:334] "Generic (PLEG): container finished" podID="3e385f5e-2610-44cd-ac51-ff0642cb47a9" containerID="aaad1146c89ce03452c1499edb200337622b1b610954c092932af6f16c0f674d" exitCode=0 Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.201256 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-8dnjd" event={"ID":"3e385f5e-2610-44cd-ac51-ff0642cb47a9","Type":"ContainerDied","Data":"aaad1146c89ce03452c1499edb200337622b1b610954c092932af6f16c0f674d"} Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.243916 4872 scope.go:117] "RemoveContainer" containerID="fa361760c26fdc8a9fca1802f8e52984ede98605e464c94607388227ac364b11" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.262130 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-p9gpz"] Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.273543 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-p9gpz"] Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.423887 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-b65567844-p9ckk"] Jan 26 09:24:02 crc kubenswrapper[4872]: E0126 09:24:02.424419 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04076889-57e8-41ad-b47e-6d8e5e23223f" containerName="neutron-db-sync" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.424442 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="04076889-57e8-41ad-b47e-6d8e5e23223f" containerName="neutron-db-sync" Jan 26 09:24:02 crc kubenswrapper[4872]: E0126 09:24:02.424464 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b" containerName="keystone-bootstrap" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.424473 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b" containerName="keystone-bootstrap" Jan 26 09:24:02 crc kubenswrapper[4872]: E0126 09:24:02.426450 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b1367db-cf5f-45c6-b0e6-b9dee6095824" containerName="dnsmasq-dns" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.426477 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b1367db-cf5f-45c6-b0e6-b9dee6095824" containerName="dnsmasq-dns" Jan 26 09:24:02 crc kubenswrapper[4872]: E0126 09:24:02.426528 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b1367db-cf5f-45c6-b0e6-b9dee6095824" containerName="init" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.426537 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b1367db-cf5f-45c6-b0e6-b9dee6095824" containerName="init" Jan 26 09:24:02 crc kubenswrapper[4872]: E0126 09:24:02.426561 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9d0b8b-dad6-4dbd-a261-77264014fc9b" containerName="placement-db-sync" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.426569 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9d0b8b-dad6-4dbd-a261-77264014fc9b" containerName="placement-db-sync" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.427079 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="04076889-57e8-41ad-b47e-6d8e5e23223f" containerName="neutron-db-sync" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.427126 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9d0b8b-dad6-4dbd-a261-77264014fc9b" containerName="placement-db-sync" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.427147 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b1367db-cf5f-45c6-b0e6-b9dee6095824" containerName="dnsmasq-dns" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.427166 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b" containerName="keystone-bootstrap" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.429336 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.433490 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.433925 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-x6c24" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.434084 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.439834 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.440139 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.449759 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b65567844-p9ckk"] Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.543732 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-internal-tls-certs\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.544049 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd7cz\" (UniqueName: \"kubernetes.io/projected/e643ca60-8ad6-4846-976d-4d2f57c70788-kube-api-access-qd7cz\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.544223 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-config-data\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.544363 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-combined-ca-bundle\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.544515 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e643ca60-8ad6-4846-976d-4d2f57c70788-logs\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.544689 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-scripts\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.544857 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-public-tls-certs\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.604481 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-769b85d4b6-8zpg4"] Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.616605 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.633569 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-769b85d4b6-8zpg4"] Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.633791 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.634098 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.634371 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.634571 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-7k5wq" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.636511 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.636873 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.650707 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-internal-tls-certs\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.651182 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd7cz\" (UniqueName: \"kubernetes.io/projected/e643ca60-8ad6-4846-976d-4d2f57c70788-kube-api-access-qd7cz\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.651377 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-config-data\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.651507 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-combined-ca-bundle\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.651681 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e643ca60-8ad6-4846-976d-4d2f57c70788-logs\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.651928 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-scripts\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.652100 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-public-tls-certs\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.653715 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e643ca60-8ad6-4846-976d-4d2f57c70788-logs\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.669334 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-public-tls-certs\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.670314 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-combined-ca-bundle\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.675600 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-internal-tls-certs\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.677576 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-config-data\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.678108 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e643ca60-8ad6-4846-976d-4d2f57c70788-scripts\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.739641 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-v8rr8"] Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.746323 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.752201 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5c4cd99bf6-6gmxf"] Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.753763 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.758580 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd7cz\" (UniqueName: \"kubernetes.io/projected/e643ca60-8ad6-4846-976d-4d2f57c70788-kube-api-access-qd7cz\") pod \"placement-b65567844-p9ckk\" (UID: \"e643ca60-8ad6-4846-976d-4d2f57c70788\") " pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.762255 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-combined-ca-bundle\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.762313 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-config-data\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.762387 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-internal-tls-certs\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.762484 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-scripts\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.762521 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-public-tls-certs\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.762571 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9kqj\" (UniqueName: \"kubernetes.io/projected/5f3edc0f-a2a4-42ee-8f69-44195674cb06-kube-api-access-d9kqj\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.762616 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-credential-keys\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.762647 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-fernet-keys\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.770321 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-nm7rl" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.770560 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.770757 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.771506 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.813470 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.822732 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-v8rr8"] Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.839768 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c4cd99bf6-6gmxf"] Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865118 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-config\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865168 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-credential-keys\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865191 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-fernet-keys\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865228 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-combined-ca-bundle\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865249 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-config-data\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865287 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-httpd-config\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865317 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865339 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-internal-tls-certs\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865398 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-combined-ca-bundle\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865423 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865445 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-config\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865462 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-ovndb-tls-certs\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865480 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bfpv\" (UniqueName: \"kubernetes.io/projected/b62ba569-ac64-47fc-b64f-284d6f51391a-kube-api-access-7bfpv\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865506 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-scripts\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865525 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-public-tls-certs\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865546 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865567 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llsh5\" (UniqueName: \"kubernetes.io/projected/5c05094d-a3b9-4016-88c1-2eabf12f3920-kube-api-access-llsh5\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865595 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9kqj\" (UniqueName: \"kubernetes.io/projected/5f3edc0f-a2a4-42ee-8f69-44195674cb06-kube-api-access-d9kqj\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.865612 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.880418 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-public-tls-certs\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.882859 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-scripts\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.885303 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-config-data\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.886715 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-credential-keys\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.887302 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-combined-ca-bundle\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.894043 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9kqj\" (UniqueName: \"kubernetes.io/projected/5f3edc0f-a2a4-42ee-8f69-44195674cb06-kube-api-access-d9kqj\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.896594 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-internal-tls-certs\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:02 crc kubenswrapper[4872]: I0126 09:24:02.897279 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5f3edc0f-a2a4-42ee-8f69-44195674cb06-fernet-keys\") pod \"keystone-769b85d4b6-8zpg4\" (UID: \"5f3edc0f-a2a4-42ee-8f69-44195674cb06\") " pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.016700 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-combined-ca-bundle\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.016783 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.016843 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-config\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.016874 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-ovndb-tls-certs\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.016902 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bfpv\" (UniqueName: \"kubernetes.io/projected/b62ba569-ac64-47fc-b64f-284d6f51391a-kube-api-access-7bfpv\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.017007 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llsh5\" (UniqueName: \"kubernetes.io/projected/5c05094d-a3b9-4016-88c1-2eabf12f3920-kube-api-access-llsh5\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.017032 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.017117 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.017185 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-config\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.017355 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-httpd-config\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.017409 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.021597 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.023626 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-config\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.024412 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.031518 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.021762 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.034269 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-combined-ca-bundle\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.037086 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.074660 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bfpv\" (UniqueName: \"kubernetes.io/projected/b62ba569-ac64-47fc-b64f-284d6f51391a-kube-api-access-7bfpv\") pod \"dnsmasq-dns-84b966f6c9-v8rr8\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.079923 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-88948b444-cjfbr"] Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.086282 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-ovndb-tls-certs\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.087179 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llsh5\" (UniqueName: \"kubernetes.io/projected/5c05094d-a3b9-4016-88c1-2eabf12f3920-kube-api-access-llsh5\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.094697 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.101547 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-88948b444-cjfbr"] Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.124884 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.126989 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-httpd-config\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.127238 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lw5r\" (UniqueName: \"kubernetes.io/projected/ab2bb608-c691-45ba-a23b-b0df661b68b7-kube-api-access-8lw5r\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.127299 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-ovndb-tls-certs\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.127329 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-config\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.127427 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-combined-ca-bundle\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.147949 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-httpd-config\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.180057 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-config\") pod \"neutron-5c4cd99bf6-6gmxf\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.232907 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lw5r\" (UniqueName: \"kubernetes.io/projected/ab2bb608-c691-45ba-a23b-b0df661b68b7-kube-api-access-8lw5r\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.233439 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-ovndb-tls-certs\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.233483 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-config\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.233554 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-combined-ca-bundle\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.233656 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-httpd-config\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.247971 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-combined-ca-bundle\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.248310 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-httpd-config\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.254350 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-config\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.266213 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-ovndb-tls-certs\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.282474 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b1367db-cf5f-45c6-b0e6-b9dee6095824" path="/var/lib/kubelet/pods/1b1367db-cf5f-45c6-b0e6-b9dee6095824/volumes" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.285562 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lw5r\" (UniqueName: \"kubernetes.io/projected/ab2bb608-c691-45ba-a23b-b0df661b68b7-kube-api-access-8lw5r\") pod \"neutron-88948b444-cjfbr\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.291890 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-8dnjd" event={"ID":"3e385f5e-2610-44cd-ac51-ff0642cb47a9","Type":"ContainerStarted","Data":"b4dd2ad3bec3723efd75a41294d3c667461da9a41770ec5d215a71d8764d270b"} Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.339290 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a7856628-f316-416e-a721-bb74270c6a02","Type":"ContainerStarted","Data":"d77a7b2f69bed239f43fb42d46d990f2e6b7804e92735447f7dfd1ca1783642e"} Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.340671 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-db-sync-8dnjd" podStartSLOduration=17.54577572 podStartE2EDuration="24.340646899s" podCreationTimestamp="2026-01-26 09:23:39 +0000 UTC" firstStartedPulling="2026-01-26 09:23:54.514304431 +0000 UTC m=+967.823144232" lastFinishedPulling="2026-01-26 09:24:01.30917561 +0000 UTC m=+974.618015411" observedRunningTime="2026-01-26 09:24:03.333387173 +0000 UTC m=+976.642226984" watchObservedRunningTime="2026-01-26 09:24:03.340646899 +0000 UTC m=+976.649486700" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.341716 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d24cd308-eaee-4bf0-be25-f4822e095b22","Type":"ContainerStarted","Data":"8e8de46c5ab125013aadce3938b5a92e6d301d8280ffd9c143b6ebeba3b60d67"} Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.434065 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.451146 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.627759 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b65567844-p9ckk"] Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.922053 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-v8rr8"] Jan 26 09:24:03 crc kubenswrapper[4872]: I0126 09:24:03.989462 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-769b85d4b6-8zpg4"] Jan 26 09:24:04 crc kubenswrapper[4872]: I0126 09:24:04.404788 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-769b85d4b6-8zpg4" event={"ID":"5f3edc0f-a2a4-42ee-8f69-44195674cb06","Type":"ContainerStarted","Data":"34af71e728c8a93836dc31d4fcb44139cc14c1149d8926249f542f68d088d257"} Jan 26 09:24:04 crc kubenswrapper[4872]: I0126 09:24:04.414361 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b65567844-p9ckk" event={"ID":"e643ca60-8ad6-4846-976d-4d2f57c70788","Type":"ContainerStarted","Data":"84c43d78ba0e5620345f148cbd150c932dec68154b1d7f09c40d499c3ca4f912"} Jan 26 09:24:04 crc kubenswrapper[4872]: I0126 09:24:04.414443 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b65567844-p9ckk" event={"ID":"e643ca60-8ad6-4846-976d-4d2f57c70788","Type":"ContainerStarted","Data":"af6a112b3dc738736122d8930edfd5f6d84906d9b9387fe162c46d649dcdbbb1"} Jan 26 09:24:04 crc kubenswrapper[4872]: I0126 09:24:04.418218 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" event={"ID":"b62ba569-ac64-47fc-b64f-284d6f51391a","Type":"ContainerStarted","Data":"536bef46f41cd5c4db426998f3b3f486cdc3a87b2fc44bdbf9cf6aa28584a7e5"} Jan 26 09:24:04 crc kubenswrapper[4872]: I0126 09:24:04.429881 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a7856628-f316-416e-a721-bb74270c6a02","Type":"ContainerStarted","Data":"54460e5c70948d40133e61aacdba15fd35cb02c9a163eac703fc6911a158ef3c"} Jan 26 09:24:04 crc kubenswrapper[4872]: I0126 09:24:04.474640 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.474615144 podStartE2EDuration="7.474615144s" podCreationTimestamp="2026-01-26 09:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:04.451221263 +0000 UTC m=+977.760061064" watchObservedRunningTime="2026-01-26 09:24:04.474615144 +0000 UTC m=+977.783454945" Jan 26 09:24:04 crc kubenswrapper[4872]: I0126 09:24:04.560714 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-88948b444-cjfbr"] Jan 26 09:24:04 crc kubenswrapper[4872]: W0126 09:24:04.584225 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab2bb608_c691_45ba_a23b_b0df661b68b7.slice/crio-46be6e2428d70aa65c521eeb69e2f2cbc70afa7c2612d424729f2d1898fa873a WatchSource:0}: Error finding container 46be6e2428d70aa65c521eeb69e2f2cbc70afa7c2612d424729f2d1898fa873a: Status 404 returned error can't find the container with id 46be6e2428d70aa65c521eeb69e2f2cbc70afa7c2612d424729f2d1898fa873a Jan 26 09:24:04 crc kubenswrapper[4872]: I0126 09:24:04.636848 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c4cd99bf6-6gmxf"] Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.386561 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c4cd99bf6-6gmxf"] Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.466707 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-795668bf-dk6bd"] Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.472706 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.480465 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.480834 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.497192 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-795668bf-dk6bd"] Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.497938 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c4cd99bf6-6gmxf" event={"ID":"5c05094d-a3b9-4016-88c1-2eabf12f3920","Type":"ContainerStarted","Data":"269f313a32b515484c32ad8ae94fc4198626212900df162c04dcd127a11599fd"} Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.497992 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c4cd99bf6-6gmxf" event={"ID":"5c05094d-a3b9-4016-88c1-2eabf12f3920","Type":"ContainerStarted","Data":"a5042937a3f42c82008c2c6a0ecec33265f7eeb6109dc01715f94a0eb6a13149"} Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.514730 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-88948b444-cjfbr" event={"ID":"ab2bb608-c691-45ba-a23b-b0df661b68b7","Type":"ContainerStarted","Data":"ef2ebf6d3a564ff492b2caf68378c294a29edd9965cc00af1dc8b28b36f98d7d"} Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.514833 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-88948b444-cjfbr" event={"ID":"ab2bb608-c691-45ba-a23b-b0df661b68b7","Type":"ContainerStarted","Data":"46be6e2428d70aa65c521eeb69e2f2cbc70afa7c2612d424729f2d1898fa873a"} Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.521258 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d24cd308-eaee-4bf0-be25-f4822e095b22","Type":"ContainerStarted","Data":"56b0345c2ca2bf94f06856c8eed362b0931457cd42f4990400a9cf7d0ab5717e"} Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.526242 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-769b85d4b6-8zpg4" event={"ID":"5f3edc0f-a2a4-42ee-8f69-44195674cb06","Type":"ContainerStarted","Data":"92773ec145b566acba10ccd093196fa64be51aafc5fe26dc1b4014c1b2ec82e6"} Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.526866 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.529023 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b65567844-p9ckk" event={"ID":"e643ca60-8ad6-4846-976d-4d2f57c70788","Type":"ContainerStarted","Data":"b764b23b77042bfbd8c029807a2fa601dcef0e86c315e48e749969f59d71150b"} Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.529342 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.531459 4872 generic.go:334] "Generic (PLEG): container finished" podID="b62ba569-ac64-47fc-b64f-284d6f51391a" containerID="57676e8b2fa37390789272cf322a5505c55467f945a44aca48d0a7be04f22338" exitCode=0 Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.533239 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" event={"ID":"b62ba569-ac64-47fc-b64f-284d6f51391a","Type":"ContainerDied","Data":"57676e8b2fa37390789272cf322a5505c55467f945a44aca48d0a7be04f22338"} Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.558434 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.55840942 podStartE2EDuration="8.55840942s" podCreationTimestamp="2026-01-26 09:23:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:05.550731964 +0000 UTC m=+978.859571765" watchObservedRunningTime="2026-01-26 09:24:05.55840942 +0000 UTC m=+978.867249221" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.595617 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-b65567844-p9ckk" podStartSLOduration=3.595586035 podStartE2EDuration="3.595586035s" podCreationTimestamp="2026-01-26 09:24:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:05.589250192 +0000 UTC m=+978.898090013" watchObservedRunningTime="2026-01-26 09:24:05.595586035 +0000 UTC m=+978.904425836" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.645505 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-httpd-config\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.645564 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-ovndb-tls-certs\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.645621 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-combined-ca-bundle\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.645703 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7cqq\" (UniqueName: \"kubernetes.io/projected/a0396e21-0a34-49b9-aa03-64c838f0da0c-kube-api-access-z7cqq\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.645731 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-config\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.645752 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-public-tls-certs\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.645945 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-internal-tls-certs\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.690644 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-769b85d4b6-8zpg4" podStartSLOduration=3.690615304 podStartE2EDuration="3.690615304s" podCreationTimestamp="2026-01-26 09:24:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:05.679132709 +0000 UTC m=+978.987972520" watchObservedRunningTime="2026-01-26 09:24:05.690615304 +0000 UTC m=+978.999455105" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.747737 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-internal-tls-certs\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.747984 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-httpd-config\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.748019 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-ovndb-tls-certs\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.748063 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-combined-ca-bundle\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.748106 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7cqq\" (UniqueName: \"kubernetes.io/projected/a0396e21-0a34-49b9-aa03-64c838f0da0c-kube-api-access-z7cqq\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.748133 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-config\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.748155 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-public-tls-certs\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.754509 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-ovndb-tls-certs\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.760025 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-internal-tls-certs\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.760040 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-config\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.760132 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-combined-ca-bundle\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.760219 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-public-tls-certs\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.767642 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a0396e21-0a34-49b9-aa03-64c838f0da0c-httpd-config\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.769381 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7cqq\" (UniqueName: \"kubernetes.io/projected/a0396e21-0a34-49b9-aa03-64c838f0da0c-kube-api-access-z7cqq\") pod \"neutron-795668bf-dk6bd\" (UID: \"a0396e21-0a34-49b9-aa03-64c838f0da0c\") " pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:05 crc kubenswrapper[4872]: I0126 09:24:05.821809 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.463436 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-795668bf-dk6bd"] Jan 26 09:24:06 crc kubenswrapper[4872]: W0126 09:24:06.484191 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0396e21_0a34_49b9_aa03_64c838f0da0c.slice/crio-b63197e4c46783255c880d51e80cac10e4ff1ec82798086481802ee88b8896ed WatchSource:0}: Error finding container b63197e4c46783255c880d51e80cac10e4ff1ec82798086481802ee88b8896ed: Status 404 returned error can't find the container with id b63197e4c46783255c880d51e80cac10e4ff1ec82798086481802ee88b8896ed Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.549124 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" event={"ID":"b62ba569-ac64-47fc-b64f-284d6f51391a","Type":"ContainerStarted","Data":"c65e3c2aab6a8002de12b441952f0f127e21e28c7b042d72727c0750261d8720"} Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.552215 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.557963 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c4cd99bf6-6gmxf" event={"ID":"5c05094d-a3b9-4016-88c1-2eabf12f3920","Type":"ContainerStarted","Data":"44cea19e0a7c47a6453c75eb2db092609a4d55ca7c7816f3b2d46300f9bf7181"} Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.558170 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c4cd99bf6-6gmxf" podUID="5c05094d-a3b9-4016-88c1-2eabf12f3920" containerName="neutron-api" containerID="cri-o://269f313a32b515484c32ad8ae94fc4198626212900df162c04dcd127a11599fd" gracePeriod=30 Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.558303 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.558361 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c4cd99bf6-6gmxf" podUID="5c05094d-a3b9-4016-88c1-2eabf12f3920" containerName="neutron-httpd" containerID="cri-o://44cea19e0a7c47a6453c75eb2db092609a4d55ca7c7816f3b2d46300f9bf7181" gracePeriod=30 Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.563563 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-795668bf-dk6bd" event={"ID":"a0396e21-0a34-49b9-aa03-64c838f0da0c","Type":"ContainerStarted","Data":"b63197e4c46783255c880d51e80cac10e4ff1ec82798086481802ee88b8896ed"} Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.570430 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-88948b444-cjfbr" event={"ID":"ab2bb608-c691-45ba-a23b-b0df661b68b7","Type":"ContainerStarted","Data":"413e857e7cc2d427c4e6329b01cc361aec394e2fab0a6588eed164991c76e0f0"} Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.570901 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.573475 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" podStartSLOduration=4.5734536519999995 podStartE2EDuration="4.573453652s" podCreationTimestamp="2026-01-26 09:24:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:06.573065112 +0000 UTC m=+979.881904923" watchObservedRunningTime="2026-01-26 09:24:06.573453652 +0000 UTC m=+979.882293453" Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.601334 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-88948b444-cjfbr" podStartSLOduration=3.601310927 podStartE2EDuration="3.601310927s" podCreationTimestamp="2026-01-26 09:24:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:06.600347812 +0000 UTC m=+979.909187613" watchObservedRunningTime="2026-01-26 09:24:06.601310927 +0000 UTC m=+979.910150728" Jan 26 09:24:06 crc kubenswrapper[4872]: I0126 09:24:06.629686 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5c4cd99bf6-6gmxf" podStartSLOduration=4.6296509740000005 podStartE2EDuration="4.629650974s" podCreationTimestamp="2026-01-26 09:24:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:06.621742431 +0000 UTC m=+979.930582252" watchObservedRunningTime="2026-01-26 09:24:06.629650974 +0000 UTC m=+979.938490775" Jan 26 09:24:07 crc kubenswrapper[4872]: I0126 09:24:07.585025 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 26 09:24:07 crc kubenswrapper[4872]: I0126 09:24:07.586202 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 26 09:24:07 crc kubenswrapper[4872]: I0126 09:24:07.592311 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-795668bf-dk6bd" event={"ID":"a0396e21-0a34-49b9-aa03-64c838f0da0c","Type":"ContainerStarted","Data":"c0b4244a2e2f0072ce9d9247a343df4abcbbbe6690adc1d973cad71e96bc4adc"} Jan 26 09:24:07 crc kubenswrapper[4872]: I0126 09:24:07.602657 4872 generic.go:334] "Generic (PLEG): container finished" podID="5c05094d-a3b9-4016-88c1-2eabf12f3920" containerID="44cea19e0a7c47a6453c75eb2db092609a4d55ca7c7816f3b2d46300f9bf7181" exitCode=0 Jan 26 09:24:07 crc kubenswrapper[4872]: I0126 09:24:07.602990 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c4cd99bf6-6gmxf" event={"ID":"5c05094d-a3b9-4016-88c1-2eabf12f3920","Type":"ContainerDied","Data":"44cea19e0a7c47a6453c75eb2db092609a4d55ca7c7816f3b2d46300f9bf7181"} Jan 26 09:24:07 crc kubenswrapper[4872]: I0126 09:24:07.604266 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:07 crc kubenswrapper[4872]: I0126 09:24:07.613701 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 26 09:24:07 crc kubenswrapper[4872]: I0126 09:24:07.613751 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 26 09:24:07 crc kubenswrapper[4872]: I0126 09:24:07.646899 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 26 09:24:07 crc kubenswrapper[4872]: I0126 09:24:07.655846 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 26 09:24:07 crc kubenswrapper[4872]: I0126 09:24:07.669542 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 26 09:24:07 crc kubenswrapper[4872]: I0126 09:24:07.680024 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 26 09:24:08 crc kubenswrapper[4872]: I0126 09:24:08.613661 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 26 09:24:08 crc kubenswrapper[4872]: I0126 09:24:08.614112 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 26 09:24:08 crc kubenswrapper[4872]: I0126 09:24:08.614122 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 26 09:24:08 crc kubenswrapper[4872]: I0126 09:24:08.614132 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 26 09:24:10 crc kubenswrapper[4872]: I0126 09:24:10.227589 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:24:10 crc kubenswrapper[4872]: I0126 09:24:10.228491 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:24:10 crc kubenswrapper[4872]: I0126 09:24:10.228565 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:24:10 crc kubenswrapper[4872]: I0126 09:24:10.229826 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ad422c6c34b0fdde5ba287cc03a049a9b7acc36305d83e2c495ecd1a880a0a0f"} pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 09:24:10 crc kubenswrapper[4872]: I0126 09:24:10.229939 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" containerID="cri-o://ad422c6c34b0fdde5ba287cc03a049a9b7acc36305d83e2c495ecd1a880a0a0f" gracePeriod=600 Jan 26 09:24:10 crc kubenswrapper[4872]: I0126 09:24:10.640788 4872 generic.go:334] "Generic (PLEG): container finished" podID="fca44d96-a000-4bf2-8283-a937b0192880" containerID="ad422c6c34b0fdde5ba287cc03a049a9b7acc36305d83e2c495ecd1a880a0a0f" exitCode=0 Jan 26 09:24:10 crc kubenswrapper[4872]: I0126 09:24:10.640980 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 09:24:10 crc kubenswrapper[4872]: I0126 09:24:10.642035 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerDied","Data":"ad422c6c34b0fdde5ba287cc03a049a9b7acc36305d83e2c495ecd1a880a0a0f"} Jan 26 09:24:10 crc kubenswrapper[4872]: I0126 09:24:10.642081 4872 scope.go:117] "RemoveContainer" containerID="df957db96f1d4f96c1bc63715f9b57ffc3271e0d6dac35d880b93d7745dd43b0" Jan 26 09:24:10 crc kubenswrapper[4872]: I0126 09:24:10.642184 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 09:24:10 crc kubenswrapper[4872]: I0126 09:24:10.675099 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 26 09:24:10 crc kubenswrapper[4872]: I0126 09:24:10.776916 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 26 09:24:11 crc kubenswrapper[4872]: I0126 09:24:11.657719 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerStarted","Data":"806f909f0f07687dd6ea8e62ed9234f07922743fa23a5604fb5f62e36f63f8b0"} Jan 26 09:24:11 crc kubenswrapper[4872]: I0126 09:24:11.666116 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e","Type":"ContainerStarted","Data":"a6d9b0e2833d69a3980ed93e23f6cb81343965d9706ed1f8745a3796fe532e86"} Jan 26 09:24:11 crc kubenswrapper[4872]: I0126 09:24:11.668485 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4cbzl" event={"ID":"3f653549-36ce-4db8-9c67-ec95b628e867","Type":"ContainerStarted","Data":"268730e4559ac67065bafe322e8335b0f075f20476100d6969b3ac265d2fedb9"} Jan 26 09:24:11 crc kubenswrapper[4872]: I0126 09:24:11.673463 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dk64w" event={"ID":"ddd0e60e-a07c-41b0-8c44-00844dfef24a","Type":"ContainerStarted","Data":"427462d4b3684a01399799fb4af8b2bcd3411a3b2bd21f4c7b8e83d49a115241"} Jan 26 09:24:11 crc kubenswrapper[4872]: I0126 09:24:11.675659 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-795668bf-dk6bd" event={"ID":"a0396e21-0a34-49b9-aa03-64c838f0da0c","Type":"ContainerStarted","Data":"718d3516811486733caf620ee929ba0acb136415403aec6ffeaf9854b6406547"} Jan 26 09:24:11 crc kubenswrapper[4872]: I0126 09:24:11.675689 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 09:24:11 crc kubenswrapper[4872]: I0126 09:24:11.675755 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 09:24:11 crc kubenswrapper[4872]: I0126 09:24:11.676430 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:11 crc kubenswrapper[4872]: I0126 09:24:11.708561 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-4cbzl" podStartSLOduration=2.884997912 podStartE2EDuration="42.708530586s" podCreationTimestamp="2026-01-26 09:23:29 +0000 UTC" firstStartedPulling="2026-01-26 09:23:30.551465372 +0000 UTC m=+943.860305173" lastFinishedPulling="2026-01-26 09:24:10.374998046 +0000 UTC m=+983.683837847" observedRunningTime="2026-01-26 09:24:11.703733312 +0000 UTC m=+985.012573113" watchObservedRunningTime="2026-01-26 09:24:11.708530586 +0000 UTC m=+985.017370407" Jan 26 09:24:11 crc kubenswrapper[4872]: I0126 09:24:11.729105 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-795668bf-dk6bd" podStartSLOduration=6.729088703 podStartE2EDuration="6.729088703s" podCreationTimestamp="2026-01-26 09:24:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:11.725944852 +0000 UTC m=+985.034784643" watchObservedRunningTime="2026-01-26 09:24:11.729088703 +0000 UTC m=+985.037928504" Jan 26 09:24:11 crc kubenswrapper[4872]: I0126 09:24:11.755915 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-dk64w" podStartSLOduration=3.28602405 podStartE2EDuration="42.755891021s" podCreationTimestamp="2026-01-26 09:23:29 +0000 UTC" firstStartedPulling="2026-01-26 09:23:30.889151586 +0000 UTC m=+944.197991387" lastFinishedPulling="2026-01-26 09:24:10.359018557 +0000 UTC m=+983.667858358" observedRunningTime="2026-01-26 09:24:11.742527628 +0000 UTC m=+985.051367439" watchObservedRunningTime="2026-01-26 09:24:11.755891021 +0000 UTC m=+985.064730822" Jan 26 09:24:12 crc kubenswrapper[4872]: I0126 09:24:12.343643 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 26 09:24:12 crc kubenswrapper[4872]: I0126 09:24:12.458693 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.131084 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.215223 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-k2frs"] Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.215601 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" podUID="eec4c66a-0959-41ef-acaa-44c65d1a9e13" containerName="dnsmasq-dns" containerID="cri-o://2d25c74d651e163d432f82fc0c94cfdf58fdc7cd4cdea162e4e71185f76541ea" gracePeriod=10 Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.749363 4872 generic.go:334] "Generic (PLEG): container finished" podID="eec4c66a-0959-41ef-acaa-44c65d1a9e13" containerID="2d25c74d651e163d432f82fc0c94cfdf58fdc7cd4cdea162e4e71185f76541ea" exitCode=0 Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.749554 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" event={"ID":"eec4c66a-0959-41ef-acaa-44c65d1a9e13","Type":"ContainerDied","Data":"2d25c74d651e163d432f82fc0c94cfdf58fdc7cd4cdea162e4e71185f76541ea"} Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.859567 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.863927 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-dns-swift-storage-0\") pod \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.864200 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-ovsdbserver-sb\") pod \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.864303 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-np59l\" (UniqueName: \"kubernetes.io/projected/eec4c66a-0959-41ef-acaa-44c65d1a9e13-kube-api-access-np59l\") pod \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.864358 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-dns-svc\") pod \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.864456 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-ovsdbserver-nb\") pod \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.864554 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-config\") pod \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\" (UID: \"eec4c66a-0959-41ef-acaa-44c65d1a9e13\") " Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.885314 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eec4c66a-0959-41ef-acaa-44c65d1a9e13-kube-api-access-np59l" (OuterVolumeSpecName: "kube-api-access-np59l") pod "eec4c66a-0959-41ef-acaa-44c65d1a9e13" (UID: "eec4c66a-0959-41ef-acaa-44c65d1a9e13"). InnerVolumeSpecName "kube-api-access-np59l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.967595 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-np59l\" (UniqueName: \"kubernetes.io/projected/eec4c66a-0959-41ef-acaa-44c65d1a9e13-kube-api-access-np59l\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:13 crc kubenswrapper[4872]: I0126 09:24:13.970727 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eec4c66a-0959-41ef-acaa-44c65d1a9e13" (UID: "eec4c66a-0959-41ef-acaa-44c65d1a9e13"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.007501 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eec4c66a-0959-41ef-acaa-44c65d1a9e13" (UID: "eec4c66a-0959-41ef-acaa-44c65d1a9e13"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.013386 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eec4c66a-0959-41ef-acaa-44c65d1a9e13" (UID: "eec4c66a-0959-41ef-acaa-44c65d1a9e13"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.017651 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eec4c66a-0959-41ef-acaa-44c65d1a9e13" (UID: "eec4c66a-0959-41ef-acaa-44c65d1a9e13"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.047791 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-config" (OuterVolumeSpecName: "config") pod "eec4c66a-0959-41ef-acaa-44c65d1a9e13" (UID: "eec4c66a-0959-41ef-acaa-44c65d1a9e13"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.069375 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.069413 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.069423 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.069434 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.069445 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eec4c66a-0959-41ef-acaa-44c65d1a9e13-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.765018 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.765042 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-k2frs" event={"ID":"eec4c66a-0959-41ef-acaa-44c65d1a9e13","Type":"ContainerDied","Data":"27d5a45559a08b8186990cd3ff92a96521155fcd7c63e66885e2b258668c57fc"} Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.766700 4872 scope.go:117] "RemoveContainer" containerID="2d25c74d651e163d432f82fc0c94cfdf58fdc7cd4cdea162e4e71185f76541ea" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.770412 4872 generic.go:334] "Generic (PLEG): container finished" podID="ddd0e60e-a07c-41b0-8c44-00844dfef24a" containerID="427462d4b3684a01399799fb4af8b2bcd3411a3b2bd21f4c7b8e83d49a115241" exitCode=0 Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.770497 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dk64w" event={"ID":"ddd0e60e-a07c-41b0-8c44-00844dfef24a","Type":"ContainerDied","Data":"427462d4b3684a01399799fb4af8b2bcd3411a3b2bd21f4c7b8e83d49a115241"} Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.819080 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-k2frs"] Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.822068 4872 scope.go:117] "RemoveContainer" containerID="fb69ec54392d85cac6272918a6198b93edae1774864b44bf93e91dbdd5783f91" Jan 26 09:24:14 crc kubenswrapper[4872]: I0126 09:24:14.825993 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-k2frs"] Jan 26 09:24:15 crc kubenswrapper[4872]: I0126 09:24:15.198125 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eec4c66a-0959-41ef-acaa-44c65d1a9e13" path="/var/lib/kubelet/pods/eec4c66a-0959-41ef-acaa-44c65d1a9e13/volumes" Jan 26 09:24:17 crc kubenswrapper[4872]: I0126 09:24:17.820284 4872 generic.go:334] "Generic (PLEG): container finished" podID="3f653549-36ce-4db8-9c67-ec95b628e867" containerID="268730e4559ac67065bafe322e8335b0f075f20476100d6969b3ac265d2fedb9" exitCode=0 Jan 26 09:24:17 crc kubenswrapper[4872]: I0126 09:24:17.820352 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4cbzl" event={"ID":"3f653549-36ce-4db8-9c67-ec95b628e867","Type":"ContainerDied","Data":"268730e4559ac67065bafe322e8335b0f075f20476100d6969b3ac265d2fedb9"} Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.798510 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dk64w" Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.839953 4872 generic.go:334] "Generic (PLEG): container finished" podID="3e385f5e-2610-44cd-ac51-ff0642cb47a9" containerID="b4dd2ad3bec3723efd75a41294d3c667461da9a41770ec5d215a71d8764d270b" exitCode=0 Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.840042 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-8dnjd" event={"ID":"3e385f5e-2610-44cd-ac51-ff0642cb47a9","Type":"ContainerDied","Data":"b4dd2ad3bec3723efd75a41294d3c667461da9a41770ec5d215a71d8764d270b"} Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.844175 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnslt\" (UniqueName: \"kubernetes.io/projected/ddd0e60e-a07c-41b0-8c44-00844dfef24a-kube-api-access-vnslt\") pod \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\" (UID: \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\") " Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.844423 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd0e60e-a07c-41b0-8c44-00844dfef24a-combined-ca-bundle\") pod \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\" (UID: \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\") " Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.844662 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ddd0e60e-a07c-41b0-8c44-00844dfef24a-db-sync-config-data\") pod \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\" (UID: \"ddd0e60e-a07c-41b0-8c44-00844dfef24a\") " Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.846868 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dk64w" event={"ID":"ddd0e60e-a07c-41b0-8c44-00844dfef24a","Type":"ContainerDied","Data":"92cff79ea1dac3cc7f82afd01745bdedaa7ae94f680e1a2de68b0f4e56c7b2b3"} Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.846913 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92cff79ea1dac3cc7f82afd01745bdedaa7ae94f680e1a2de68b0f4e56c7b2b3" Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.846919 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dk64w" Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.852505 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddd0e60e-a07c-41b0-8c44-00844dfef24a-kube-api-access-vnslt" (OuterVolumeSpecName: "kube-api-access-vnslt") pod "ddd0e60e-a07c-41b0-8c44-00844dfef24a" (UID: "ddd0e60e-a07c-41b0-8c44-00844dfef24a"). InnerVolumeSpecName "kube-api-access-vnslt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.855083 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddd0e60e-a07c-41b0-8c44-00844dfef24a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ddd0e60e-a07c-41b0-8c44-00844dfef24a" (UID: "ddd0e60e-a07c-41b0-8c44-00844dfef24a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.879859 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddd0e60e-a07c-41b0-8c44-00844dfef24a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ddd0e60e-a07c-41b0-8c44-00844dfef24a" (UID: "ddd0e60e-a07c-41b0-8c44-00844dfef24a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.947875 4872 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ddd0e60e-a07c-41b0-8c44-00844dfef24a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.947917 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnslt\" (UniqueName: \"kubernetes.io/projected/ddd0e60e-a07c-41b0-8c44-00844dfef24a-kube-api-access-vnslt\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:18 crc kubenswrapper[4872]: I0126 09:24:18.947934 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddd0e60e-a07c-41b0-8c44-00844dfef24a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:19 crc kubenswrapper[4872]: I0126 09:24:19.894683 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-4cbzl" event={"ID":"3f653549-36ce-4db8-9c67-ec95b628e867","Type":"ContainerDied","Data":"e7a09dbdbfd182e1d55d19f9e397be1330815a9901eb3d38bddf1feaecf09402"} Jan 26 09:24:19 crc kubenswrapper[4872]: I0126 09:24:19.895236 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7a09dbdbfd182e1d55d19f9e397be1330815a9901eb3d38bddf1feaecf09402" Jan 26 09:24:19 crc kubenswrapper[4872]: I0126 09:24:19.948117 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:24:19 crc kubenswrapper[4872]: I0126 09:24:19.978774 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f653549-36ce-4db8-9c67-ec95b628e867-etc-machine-id\") pod \"3f653549-36ce-4db8-9c67-ec95b628e867\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " Jan 26 09:24:19 crc kubenswrapper[4872]: I0126 09:24:19.979330 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-scripts\") pod \"3f653549-36ce-4db8-9c67-ec95b628e867\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " Jan 26 09:24:19 crc kubenswrapper[4872]: I0126 09:24:19.979397 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-combined-ca-bundle\") pod \"3f653549-36ce-4db8-9c67-ec95b628e867\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " Jan 26 09:24:19 crc kubenswrapper[4872]: I0126 09:24:19.979493 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-config-data\") pod \"3f653549-36ce-4db8-9c67-ec95b628e867\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " Jan 26 09:24:19 crc kubenswrapper[4872]: I0126 09:24:19.979542 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-db-sync-config-data\") pod \"3f653549-36ce-4db8-9c67-ec95b628e867\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " Jan 26 09:24:19 crc kubenswrapper[4872]: I0126 09:24:19.979571 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvmw7\" (UniqueName: \"kubernetes.io/projected/3f653549-36ce-4db8-9c67-ec95b628e867-kube-api-access-kvmw7\") pod \"3f653549-36ce-4db8-9c67-ec95b628e867\" (UID: \"3f653549-36ce-4db8-9c67-ec95b628e867\") " Jan 26 09:24:19 crc kubenswrapper[4872]: I0126 09:24:19.978865 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f653549-36ce-4db8-9c67-ec95b628e867-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3f653549-36ce-4db8-9c67-ec95b628e867" (UID: "3f653549-36ce-4db8-9c67-ec95b628e867"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:24:19 crc kubenswrapper[4872]: I0126 09:24:19.984359 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-scripts" (OuterVolumeSpecName: "scripts") pod "3f653549-36ce-4db8-9c67-ec95b628e867" (UID: "3f653549-36ce-4db8-9c67-ec95b628e867"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:19 crc kubenswrapper[4872]: I0126 09:24:19.991509 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f653549-36ce-4db8-9c67-ec95b628e867-kube-api-access-kvmw7" (OuterVolumeSpecName: "kube-api-access-kvmw7") pod "3f653549-36ce-4db8-9c67-ec95b628e867" (UID: "3f653549-36ce-4db8-9c67-ec95b628e867"). InnerVolumeSpecName "kube-api-access-kvmw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:19.998082 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3f653549-36ce-4db8-9c67-ec95b628e867" (UID: "3f653549-36ce-4db8-9c67-ec95b628e867"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.087455 4872 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.087497 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvmw7\" (UniqueName: \"kubernetes.io/projected/3f653549-36ce-4db8-9c67-ec95b628e867-kube-api-access-kvmw7\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.087513 4872 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3f653549-36ce-4db8-9c67-ec95b628e867-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.087522 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.140789 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f653549-36ce-4db8-9c67-ec95b628e867" (UID: "3f653549-36ce-4db8-9c67-ec95b628e867"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.187791 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5f44467695-j8db6"] Jan 26 09:24:20 crc kubenswrapper[4872]: E0126 09:24:20.188711 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec4c66a-0959-41ef-acaa-44c65d1a9e13" containerName="init" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.188735 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec4c66a-0959-41ef-acaa-44c65d1a9e13" containerName="init" Jan 26 09:24:20 crc kubenswrapper[4872]: E0126 09:24:20.188750 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec4c66a-0959-41ef-acaa-44c65d1a9e13" containerName="dnsmasq-dns" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.188757 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec4c66a-0959-41ef-acaa-44c65d1a9e13" containerName="dnsmasq-dns" Jan 26 09:24:20 crc kubenswrapper[4872]: E0126 09:24:20.188771 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddd0e60e-a07c-41b0-8c44-00844dfef24a" containerName="barbican-db-sync" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.188779 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddd0e60e-a07c-41b0-8c44-00844dfef24a" containerName="barbican-db-sync" Jan 26 09:24:20 crc kubenswrapper[4872]: E0126 09:24:20.188826 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f653549-36ce-4db8-9c67-ec95b628e867" containerName="cinder-db-sync" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.188832 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f653549-36ce-4db8-9c67-ec95b628e867" containerName="cinder-db-sync" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.189023 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f653549-36ce-4db8-9c67-ec95b628e867" containerName="cinder-db-sync" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.189046 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddd0e60e-a07c-41b0-8c44-00844dfef24a" containerName="barbican-db-sync" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.189068 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec4c66a-0959-41ef-acaa-44c65d1a9e13" containerName="dnsmasq-dns" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.190081 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.207876 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.208587 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-zrmml" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.220545 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.221055 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.240854 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5f44467695-j8db6"] Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.248196 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-config-data" (OuterVolumeSpecName: "config-data") pod "3f653549-36ce-4db8-9c67-ec95b628e867" (UID: "3f653549-36ce-4db8-9c67-ec95b628e867"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.286098 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-664886958b-t96tw"] Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.292110 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.295317 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.310404 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2hsh\" (UniqueName: \"kubernetes.io/projected/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-kube-api-access-l2hsh\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.310551 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-config-data\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.310580 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-logs\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.310608 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-config-data-custom\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.310627 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-combined-ca-bundle\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.310692 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f653549-36ce-4db8-9c67-ec95b628e867-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.315931 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-664886958b-t96tw"] Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.347243 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-vbpqn"] Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.349341 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.361847 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-vbpqn"] Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.412096 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-config-data-custom\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.412156 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6532ea83-7260-4883-8cdd-236c90c5c90a-config-data\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.412185 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-combined-ca-bundle\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.412207 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6532ea83-7260-4883-8cdd-236c90c5c90a-combined-ca-bundle\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.412247 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2hsh\" (UniqueName: \"kubernetes.io/projected/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-kube-api-access-l2hsh\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.412431 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6532ea83-7260-4883-8cdd-236c90c5c90a-logs\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.412493 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6532ea83-7260-4883-8cdd-236c90c5c90a-config-data-custom\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.412517 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7752\" (UniqueName: \"kubernetes.io/projected/6532ea83-7260-4883-8cdd-236c90c5c90a-kube-api-access-l7752\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.412547 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-config-data\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.412575 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-logs\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.414158 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-logs\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.421259 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-config-data-custom\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.421929 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-78b87b9446-n4fxp"] Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.425737 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.427906 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-config-data\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.431423 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.431434 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-combined-ca-bundle\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.432969 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-78b87b9446-n4fxp"] Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.446070 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2hsh\" (UniqueName: \"kubernetes.io/projected/1dc34b32-f172-4b29-8bd1-a9f7a86d0996-kube-api-access-l2hsh\") pod \"barbican-worker-5f44467695-j8db6\" (UID: \"1dc34b32-f172-4b29-8bd1-a9f7a86d0996\") " pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.454452 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515240 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6532ea83-7260-4883-8cdd-236c90c5c90a-combined-ca-bundle\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515314 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515338 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515362 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515397 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data-custom\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515434 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-config\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515467 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515502 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6532ea83-7260-4883-8cdd-236c90c5c90a-logs\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515523 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/421c2c04-702f-403f-a6b0-fa784dd1af3f-logs\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515549 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515584 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tflrm\" (UniqueName: \"kubernetes.io/projected/421c2c04-702f-403f-a6b0-fa784dd1af3f-kube-api-access-tflrm\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515617 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6532ea83-7260-4883-8cdd-236c90c5c90a-config-data-custom\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515637 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7752\" (UniqueName: \"kubernetes.io/projected/6532ea83-7260-4883-8cdd-236c90c5c90a-kube-api-access-l7752\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515657 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-combined-ca-bundle\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515702 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bps42\" (UniqueName: \"kubernetes.io/projected/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-kube-api-access-bps42\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.515730 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6532ea83-7260-4883-8cdd-236c90c5c90a-config-data\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.516696 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6532ea83-7260-4883-8cdd-236c90c5c90a-logs\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.520451 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6532ea83-7260-4883-8cdd-236c90c5c90a-config-data\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.520672 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6532ea83-7260-4883-8cdd-236c90c5c90a-combined-ca-bundle\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.533621 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6532ea83-7260-4883-8cdd-236c90c5c90a-config-data-custom\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.544426 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7752\" (UniqueName: \"kubernetes.io/projected/6532ea83-7260-4883-8cdd-236c90c5c90a-kube-api-access-l7752\") pod \"barbican-keystone-listener-664886958b-t96tw\" (UID: \"6532ea83-7260-4883-8cdd-236c90c5c90a\") " pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.557900 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5f44467695-j8db6" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.619655 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75gnr\" (UniqueName: \"kubernetes.io/projected/3e385f5e-2610-44cd-ac51-ff0642cb47a9-kube-api-access-75gnr\") pod \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.619827 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3e385f5e-2610-44cd-ac51-ff0642cb47a9-config-data-merged\") pod \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.619930 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-scripts\") pod \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.619980 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-config-data\") pod \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.620111 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/3e385f5e-2610-44cd-ac51-ff0642cb47a9-etc-podinfo\") pod \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.621617 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-combined-ca-bundle\") pod \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\" (UID: \"3e385f5e-2610-44cd-ac51-ff0642cb47a9\") " Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.622031 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bps42\" (UniqueName: \"kubernetes.io/projected/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-kube-api-access-bps42\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.622122 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.622149 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.622184 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.622224 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data-custom\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.622268 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-config\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.622509 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.622597 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/421c2c04-702f-403f-a6b0-fa784dd1af3f-logs\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.623306 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.625201 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e385f5e-2610-44cd-ac51-ff0642cb47a9-kube-api-access-75gnr" (OuterVolumeSpecName: "kube-api-access-75gnr") pod "3e385f5e-2610-44cd-ac51-ff0642cb47a9" (UID: "3e385f5e-2610-44cd-ac51-ff0642cb47a9"). InnerVolumeSpecName "kube-api-access-75gnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.625851 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/421c2c04-702f-403f-a6b0-fa784dd1af3f-logs\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.626397 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3e385f5e-2610-44cd-ac51-ff0642cb47a9-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "3e385f5e-2610-44cd-ac51-ff0642cb47a9" (UID: "3e385f5e-2610-44cd-ac51-ff0642cb47a9"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.626577 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-scripts" (OuterVolumeSpecName: "scripts") pod "3e385f5e-2610-44cd-ac51-ff0642cb47a9" (UID: "3e385f5e-2610-44cd-ac51-ff0642cb47a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.628059 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.628156 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tflrm\" (UniqueName: \"kubernetes.io/projected/421c2c04-702f-403f-a6b0-fa784dd1af3f-kube-api-access-tflrm\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.628259 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-combined-ca-bundle\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.628646 4872 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/3e385f5e-2610-44cd-ac51-ff0642cb47a9-etc-podinfo\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.628690 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75gnr\" (UniqueName: \"kubernetes.io/projected/3e385f5e-2610-44cd-ac51-ff0642cb47a9-kube-api-access-75gnr\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.628723 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.630184 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e385f5e-2610-44cd-ac51-ff0642cb47a9-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "3e385f5e-2610-44cd-ac51-ff0642cb47a9" (UID: "3e385f5e-2610-44cd-ac51-ff0642cb47a9"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.632230 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.633093 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.635392 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data-custom\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.635836 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.635901 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-config\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.636266 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.638768 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-combined-ca-bundle\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.652608 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bps42\" (UniqueName: \"kubernetes.io/projected/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-kube-api-access-bps42\") pod \"dnsmasq-dns-75c8ddd69c-vbpqn\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.656223 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tflrm\" (UniqueName: \"kubernetes.io/projected/421c2c04-702f-403f-a6b0-fa784dd1af3f-kube-api-access-tflrm\") pod \"barbican-api-78b87b9446-n4fxp\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.661745 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-config-data" (OuterVolumeSpecName: "config-data") pod "3e385f5e-2610-44cd-ac51-ff0642cb47a9" (UID: "3e385f5e-2610-44cd-ac51-ff0642cb47a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.662940 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-664886958b-t96tw" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.695720 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.700460 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e385f5e-2610-44cd-ac51-ff0642cb47a9" (UID: "3e385f5e-2610-44cd-ac51-ff0642cb47a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.732311 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.732358 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e385f5e-2610-44cd-ac51-ff0642cb47a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.732377 4872 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/3e385f5e-2610-44cd-ac51-ff0642cb47a9-config-data-merged\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.767855 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.948557 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-8dnjd" event={"ID":"3e385f5e-2610-44cd-ac51-ff0642cb47a9","Type":"ContainerDied","Data":"d15a1e6f007a8c6847574b0720848d9f2996d30a3050e89a09a1953bd95ca41a"} Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.949005 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d15a1e6f007a8c6847574b0720848d9f2996d30a3050e89a09a1953bd95ca41a" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.949099 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-8dnjd" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.956345 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-4cbzl" Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.956724 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="ceilometer-central-agent" containerID="cri-o://6aa5198368a8ddf875c04c2c19c0ee31224b4ec54781dc7d825df99106c52642" gracePeriod=30 Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.957010 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="ceilometer-notification-agent" containerID="cri-o://02b520f81790fe052d10b40bbca3e30c9b45bb327c1aab132b4362a8bbda7d52" gracePeriod=30 Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.957035 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="sg-core" containerID="cri-o://a6d9b0e2833d69a3980ed93e23f6cb81343965d9706ed1f8745a3796fe532e86" gracePeriod=30 Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.957169 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="proxy-httpd" containerID="cri-o://bd643af7e6e14e0c065addf247a3ba8e0f02b3186d10c0d31402c4fa4a6ddb22" gracePeriod=30 Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.957198 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e","Type":"ContainerStarted","Data":"bd643af7e6e14e0c065addf247a3ba8e0f02b3186d10c0d31402c4fa4a6ddb22"} Jan 26 09:24:20 crc kubenswrapper[4872]: I0126 09:24:20.957240 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.017543 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.878334527 podStartE2EDuration="52.017516906s" podCreationTimestamp="2026-01-26 09:23:29 +0000 UTC" firstStartedPulling="2026-01-26 09:23:30.800026711 +0000 UTC m=+944.108866512" lastFinishedPulling="2026-01-26 09:24:19.93920908 +0000 UTC m=+993.248048891" observedRunningTime="2026-01-26 09:24:20.986850408 +0000 UTC m=+994.295690219" watchObservedRunningTime="2026-01-26 09:24:21.017516906 +0000 UTC m=+994.326356707" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.065263 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5f44467695-j8db6"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.315525 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-db-create-mhtml"] Jan 26 09:24:21 crc kubenswrapper[4872]: E0126 09:24:21.316625 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e385f5e-2610-44cd-ac51-ff0642cb47a9" containerName="ironic-db-sync" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.316646 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e385f5e-2610-44cd-ac51-ff0642cb47a9" containerName="ironic-db-sync" Jan 26 09:24:21 crc kubenswrapper[4872]: E0126 09:24:21.316661 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e385f5e-2610-44cd-ac51-ff0642cb47a9" containerName="init" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.316668 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e385f5e-2610-44cd-ac51-ff0642cb47a9" containerName="init" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.316895 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e385f5e-2610-44cd-ac51-ff0642cb47a9" containerName="ironic-db-sync" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.317525 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-mhtml" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.346759 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.348438 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.352268 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.365546 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.365831 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.366003 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-gqlrs" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.370322 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-create-mhtml"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.449821 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.518390 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.518490 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-scripts\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.518658 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.518713 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e205129d-e73c-4c6f-9110-788265aaada0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.518819 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2-operator-scripts\") pod \"ironic-inspector-db-create-mhtml\" (UID: \"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2\") " pod="openstack/ironic-inspector-db-create-mhtml" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.518928 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcn24\" (UniqueName: \"kubernetes.io/projected/e205129d-e73c-4c6f-9110-788265aaada0-kube-api-access-fcn24\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.518976 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-config-data\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.519025 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpsbf\" (UniqueName: \"kubernetes.io/projected/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2-kube-api-access-jpsbf\") pod \"ironic-inspector-db-create-mhtml\" (UID: \"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2\") " pod="openstack/ironic-inspector-db-create-mhtml" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.621833 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2-operator-scripts\") pod \"ironic-inspector-db-create-mhtml\" (UID: \"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2\") " pod="openstack/ironic-inspector-db-create-mhtml" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.621908 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcn24\" (UniqueName: \"kubernetes.io/projected/e205129d-e73c-4c6f-9110-788265aaada0-kube-api-access-fcn24\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.621940 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-config-data\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.621969 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpsbf\" (UniqueName: \"kubernetes.io/projected/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2-kube-api-access-jpsbf\") pod \"ironic-inspector-db-create-mhtml\" (UID: \"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2\") " pod="openstack/ironic-inspector-db-create-mhtml" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.622004 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.622031 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-scripts\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.622084 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.622116 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e205129d-e73c-4c6f-9110-788265aaada0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.622228 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e205129d-e73c-4c6f-9110-788265aaada0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.644856 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2-operator-scripts\") pod \"ironic-inspector-db-create-mhtml\" (UID: \"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2\") " pod="openstack/ironic-inspector-db-create-mhtml" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.657958 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-config-data\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.670695 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.674409 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-scripts\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.686908 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.699571 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpsbf\" (UniqueName: \"kubernetes.io/projected/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2-kube-api-access-jpsbf\") pod \"ironic-inspector-db-create-mhtml\" (UID: \"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2\") " pod="openstack/ironic-inspector-db-create-mhtml" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.700710 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcn24\" (UniqueName: \"kubernetes.io/projected/e205129d-e73c-4c6f-9110-788265aaada0-kube-api-access-fcn24\") pod \"cinder-scheduler-0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.718291 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-664886958b-t96tw"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.726638 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.728424 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.736337 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-dockercfg-6kfkt" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.738060 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-neutron-agent-config-data" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.758905 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.765922 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-vbpqn"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.824864 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.827060 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f721aeee-7daf-413d-885b-5c146881eb99-combined-ca-bundle\") pod \"ironic-neutron-agent-5b47d6fc7b-7wkw2\" (UID: \"f721aeee-7daf-413d-885b-5c146881eb99\") " pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.827141 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmbhh\" (UniqueName: \"kubernetes.io/projected/f721aeee-7daf-413d-885b-5c146881eb99-kube-api-access-fmbhh\") pod \"ironic-neutron-agent-5b47d6fc7b-7wkw2\" (UID: \"f721aeee-7daf-413d-885b-5c146881eb99\") " pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.827224 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f721aeee-7daf-413d-885b-5c146881eb99-config\") pod \"ironic-neutron-agent-5b47d6fc7b-7wkw2\" (UID: \"f721aeee-7daf-413d-885b-5c146881eb99\") " pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.863693 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-b7e6-account-create-update-mdh5l"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.869144 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.883317 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-db-secret" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.893153 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-b7e6-account-create-update-mdh5l"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.904585 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-78b87b9446-n4fxp"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.917398 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-vbpqn"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.927072 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qnrbz"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.929028 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.935292 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f721aeee-7daf-413d-885b-5c146881eb99-config\") pod \"ironic-neutron-agent-5b47d6fc7b-7wkw2\" (UID: \"f721aeee-7daf-413d-885b-5c146881eb99\") " pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.936660 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f721aeee-7daf-413d-885b-5c146881eb99-combined-ca-bundle\") pod \"ironic-neutron-agent-5b47d6fc7b-7wkw2\" (UID: \"f721aeee-7daf-413d-885b-5c146881eb99\") " pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.936729 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a7aa3be-d2ee-415a-87ac-bb65176dee59-operator-scripts\") pod \"ironic-inspector-b7e6-account-create-update-mdh5l\" (UID: \"8a7aa3be-d2ee-415a-87ac-bb65176dee59\") " pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.936915 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qnrbz"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.939255 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmbhh\" (UniqueName: \"kubernetes.io/projected/f721aeee-7daf-413d-885b-5c146881eb99-kube-api-access-fmbhh\") pod \"ironic-neutron-agent-5b47d6fc7b-7wkw2\" (UID: \"f721aeee-7daf-413d-885b-5c146881eb99\") " pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.939336 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk96c\" (UniqueName: \"kubernetes.io/projected/8a7aa3be-d2ee-415a-87ac-bb65176dee59-kube-api-access-lk96c\") pod \"ironic-inspector-b7e6-account-create-update-mdh5l\" (UID: \"8a7aa3be-d2ee-415a-87ac-bb65176dee59\") " pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.945966 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f721aeee-7daf-413d-885b-5c146881eb99-combined-ca-bundle\") pod \"ironic-neutron-agent-5b47d6fc7b-7wkw2\" (UID: \"f721aeee-7daf-413d-885b-5c146881eb99\") " pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.947008 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-bc6dfb648-7gm8x"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.948852 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f721aeee-7daf-413d-885b-5c146881eb99-config\") pod \"ironic-neutron-agent-5b47d6fc7b-7wkw2\" (UID: \"f721aeee-7daf-413d-885b-5c146881eb99\") " pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.949952 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.955768 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-mhtml" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.963650 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-config-data" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.963921 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-api-scripts" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.964055 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-api-config-data" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.964227 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.978398 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-bc6dfb648-7gm8x"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.992634 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmbhh\" (UniqueName: \"kubernetes.io/projected/f721aeee-7daf-413d-885b-5c146881eb99-kube-api-access-fmbhh\") pod \"ironic-neutron-agent-5b47d6fc7b-7wkw2\" (UID: \"f721aeee-7daf-413d-885b-5c146881eb99\") " pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.993855 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 26 09:24:21 crc kubenswrapper[4872]: I0126 09:24:21.996025 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:21.997906 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78b87b9446-n4fxp" event={"ID":"421c2c04-702f-403f-a6b0-fa784dd1af3f","Type":"ContainerStarted","Data":"fdb297fa0288c7141bb3c4724a52271c2951338a799be5ea60540e97b40f5269"} Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:21.999640 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.001574 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" event={"ID":"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb","Type":"ContainerStarted","Data":"5b023608b1258b6dd6f994df3ed12e940e546c9ff22309d219a290d87d043f4e"} Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.003668 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.008510 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-664886958b-t96tw" event={"ID":"6532ea83-7260-4883-8cdd-236c90c5c90a","Type":"ContainerStarted","Data":"2f971cd392b5a7f96aae670fab90239cc0a3db4f9617a1072cbbc45df6364242"} Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.025542 4872 generic.go:334] "Generic (PLEG): container finished" podID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerID="bd643af7e6e14e0c065addf247a3ba8e0f02b3186d10c0d31402c4fa4a6ddb22" exitCode=0 Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.025578 4872 generic.go:334] "Generic (PLEG): container finished" podID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerID="a6d9b0e2833d69a3980ed93e23f6cb81343965d9706ed1f8745a3796fe532e86" exitCode=2 Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.025593 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e","Type":"ContainerDied","Data":"bd643af7e6e14e0c065addf247a3ba8e0f02b3186d10c0d31402c4fa4a6ddb22"} Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.025654 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e","Type":"ContainerDied","Data":"a6d9b0e2833d69a3980ed93e23f6cb81343965d9706ed1f8745a3796fe532e86"} Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.039060 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5f44467695-j8db6" event={"ID":"1dc34b32-f172-4b29-8bd1-a9f7a86d0996","Type":"ContainerStarted","Data":"430672c13375dd1c24435841384bd2cd28021528979625da82b4fa8f828de034"} Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.042744 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4lz4\" (UniqueName: \"kubernetes.io/projected/2b1a1c03-c8bb-44ab-b607-722109c01b5c-kube-api-access-s4lz4\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.042885 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data-custom\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.042928 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.042976 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a7aa3be-d2ee-415a-87ac-bb65176dee59-operator-scripts\") pod \"ironic-inspector-b7e6-account-create-update-mdh5l\" (UID: \"8a7aa3be-d2ee-415a-87ac-bb65176dee59\") " pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.043013 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-scripts\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.043040 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data-merged\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.043071 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/2b1a1c03-c8bb-44ab-b607-722109c01b5c-etc-podinfo\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.043101 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk96c\" (UniqueName: \"kubernetes.io/projected/8a7aa3be-d2ee-415a-87ac-bb65176dee59-kube-api-access-lk96c\") pod \"ironic-inspector-b7e6-account-create-update-mdh5l\" (UID: \"8a7aa3be-d2ee-415a-87ac-bb65176dee59\") " pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.043128 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.043148 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grtb4\" (UniqueName: \"kubernetes.io/projected/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-kube-api-access-grtb4\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.043171 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-dns-svc\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.043206 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.043225 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-config\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.043245 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-combined-ca-bundle\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.043261 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b1a1c03-c8bb-44ab-b607-722109c01b5c-logs\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.043291 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.044480 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a7aa3be-d2ee-415a-87ac-bb65176dee59-operator-scripts\") pod \"ironic-inspector-b7e6-account-create-update-mdh5l\" (UID: \"8a7aa3be-d2ee-415a-87ac-bb65176dee59\") " pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.072937 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk96c\" (UniqueName: \"kubernetes.io/projected/8a7aa3be-d2ee-415a-87ac-bb65176dee59-kube-api-access-lk96c\") pod \"ironic-inspector-b7e6-account-create-update-mdh5l\" (UID: \"8a7aa3be-d2ee-415a-87ac-bb65176dee59\") " pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.073260 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145177 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145236 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4lz4\" (UniqueName: \"kubernetes.io/projected/2b1a1c03-c8bb-44ab-b607-722109c01b5c-kube-api-access-s4lz4\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145274 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data-custom\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145315 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-config-data\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145336 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145355 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac539301-c457-4332-ad69-87e68d920938-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145390 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145417 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-scripts\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145438 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac539301-c457-4332-ad69-87e68d920938-logs\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145455 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data-merged\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145482 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/2b1a1c03-c8bb-44ab-b607-722109c01b5c-etc-podinfo\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145518 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145538 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grtb4\" (UniqueName: \"kubernetes.io/projected/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-kube-api-access-grtb4\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145553 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-scripts\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145576 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-dns-svc\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145597 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pns7\" (UniqueName: \"kubernetes.io/projected/ac539301-c457-4332-ad69-87e68d920938-kube-api-access-8pns7\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145627 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145645 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-config\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145666 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-combined-ca-bundle\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145684 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b1a1c03-c8bb-44ab-b607-722109c01b5c-logs\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.145703 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-config-data-custom\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.148105 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.148334 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-dns-svc\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.148639 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.149001 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.149637 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-config\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.151244 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b1a1c03-c8bb-44ab-b607-722109c01b5c-logs\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.151285 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data-merged\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.154491 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-scripts\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.155011 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data-custom\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.155063 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.157414 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-combined-ca-bundle\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.158830 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/2b1a1c03-c8bb-44ab-b607-722109c01b5c-etc-podinfo\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.164474 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grtb4\" (UniqueName: \"kubernetes.io/projected/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-kube-api-access-grtb4\") pod \"dnsmasq-dns-5784cf869f-qnrbz\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.165383 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4lz4\" (UniqueName: \"kubernetes.io/projected/2b1a1c03-c8bb-44ab-b607-722109c01b5c-kube-api-access-s4lz4\") pod \"ironic-bc6dfb648-7gm8x\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.228919 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.247234 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-config-data-custom\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.247323 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-config-data\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.247352 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac539301-c457-4332-ad69-87e68d920938-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.247392 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.247420 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac539301-c457-4332-ad69-87e68d920938-logs\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.247465 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-scripts\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.247491 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pns7\" (UniqueName: \"kubernetes.io/projected/ac539301-c457-4332-ad69-87e68d920938-kube-api-access-8pns7\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.249796 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac539301-c457-4332-ad69-87e68d920938-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.254303 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac539301-c457-4332-ad69-87e68d920938-logs\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.270088 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.270624 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-scripts\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.271694 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.272793 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-config-data-custom\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.282784 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-config-data\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.302696 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pns7\" (UniqueName: \"kubernetes.io/projected/ac539301-c457-4332-ad69-87e68d920938-kube-api-access-8pns7\") pod \"cinder-api-0\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.306314 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.315491 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.565573 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.693281 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-conductor-0"] Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.703488 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.707071 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-conductor-scripts" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.707279 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-conductor-config-data" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.719904 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-conductor-0"] Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.740494 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-create-mhtml"] Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.805864 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2"] Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.871051 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e135958c-61c2-4199-94b3-2a8e6623310c-scripts\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.871163 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/e135958c-61c2-4199-94b3-2a8e6623310c-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.871208 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e135958c-61c2-4199-94b3-2a8e6623310c-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.871235 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk84g\" (UniqueName: \"kubernetes.io/projected/e135958c-61c2-4199-94b3-2a8e6623310c-kube-api-access-jk84g\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.871272 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e135958c-61c2-4199-94b3-2a8e6623310c-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.871350 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e135958c-61c2-4199-94b3-2a8e6623310c-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.871391 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.871474 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e135958c-61c2-4199-94b3-2a8e6623310c-config-data\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.976518 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.976625 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e135958c-61c2-4199-94b3-2a8e6623310c-config-data\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.976672 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e135958c-61c2-4199-94b3-2a8e6623310c-scripts\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.976706 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/e135958c-61c2-4199-94b3-2a8e6623310c-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.976740 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e135958c-61c2-4199-94b3-2a8e6623310c-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.976759 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk84g\" (UniqueName: \"kubernetes.io/projected/e135958c-61c2-4199-94b3-2a8e6623310c-kube-api-access-jk84g\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.976783 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e135958c-61c2-4199-94b3-2a8e6623310c-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.976886 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e135958c-61c2-4199-94b3-2a8e6623310c-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:22 crc kubenswrapper[4872]: I0126 09:24:22.977364 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e135958c-61c2-4199-94b3-2a8e6623310c-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.018511 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/e135958c-61c2-4199-94b3-2a8e6623310c-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.018648 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ironic-conductor-0" Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.018285 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e135958c-61c2-4199-94b3-2a8e6623310c-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.021660 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e135958c-61c2-4199-94b3-2a8e6623310c-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.023435 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e135958c-61c2-4199-94b3-2a8e6623310c-config-data\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.030345 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e135958c-61c2-4199-94b3-2a8e6623310c-scripts\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.036747 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qnrbz"] Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.131833 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk84g\" (UniqueName: \"kubernetes.io/projected/e135958c-61c2-4199-94b3-2a8e6623310c-kube-api-access-jk84g\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.175033 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-mhtml" event={"ID":"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2","Type":"ContainerStarted","Data":"4a317b5da5825148cbdabe98b670b45db1ce3c594aded7a03da572c242eb7ddb"} Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.207038 4872 generic.go:334] "Generic (PLEG): container finished" podID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerID="6aa5198368a8ddf875c04c2c19c0ee31224b4ec54781dc7d825df99106c52642" exitCode=0 Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.207169 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e","Type":"ContainerDied","Data":"6aa5198368a8ddf875c04c2c19c0ee31224b4ec54781dc7d825df99106c52642"} Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.210810 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-b7e6-account-create-update-mdh5l"] Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.223274 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78b87b9446-n4fxp" event={"ID":"421c2c04-702f-403f-a6b0-fa784dd1af3f","Type":"ContainerStarted","Data":"34890775e3d16135eed7f885879c0a9df37767fe57fd212923ef600d38b6a25d"} Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.231192 4872 generic.go:334] "Generic (PLEG): container finished" podID="5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb" containerID="c30824a7e633a670b986168841da9f3c10f88d1edbce9977dababd548e174e1e" exitCode=0 Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.231271 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" event={"ID":"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb","Type":"ContainerDied","Data":"c30824a7e633a670b986168841da9f3c10f88d1edbce9977dababd548e174e1e"} Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.240409 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.246986 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e205129d-e73c-4c6f-9110-788265aaada0","Type":"ContainerStarted","Data":"db4cca304f5ed0fa693cbcadd873d0e0eefe9b9f0690ef0bcb5befbe168d4bfa"} Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.287999 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ironic-conductor-0\" (UID: \"e135958c-61c2-4199-94b3-2a8e6623310c\") " pod="openstack/ironic-conductor-0" Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.330695 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-conductor-0" Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.423347 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-bc6dfb648-7gm8x"] Jan 26 09:24:23 crc kubenswrapper[4872]: I0126 09:24:23.939186 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 26 09:24:24 crc kubenswrapper[4872]: W0126 09:24:24.201570 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a7aa3be_d2ee_415a_87ac_bb65176dee59.slice/crio-0f7a20129d84c55ad86f98c7e805d29ffc41c4cb7e52c954861c6e8c1caa15ad WatchSource:0}: Error finding container 0f7a20129d84c55ad86f98c7e805d29ffc41c4cb7e52c954861c6e8c1caa15ad: Status 404 returned error can't find the container with id 0f7a20129d84c55ad86f98c7e805d29ffc41c4cb7e52c954861c6e8c1caa15ad Jan 26 09:24:24 crc kubenswrapper[4872]: W0126 09:24:24.224354 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf721aeee_7daf_413d_885b_5c146881eb99.slice/crio-7531fc9fa543edd2f37229ea844e5959dcdbc9b72fa510c717ff49e74e09dd38 WatchSource:0}: Error finding container 7531fc9fa543edd2f37229ea844e5959dcdbc9b72fa510c717ff49e74e09dd38: Status 404 returned error can't find the container with id 7531fc9fa543edd2f37229ea844e5959dcdbc9b72fa510c717ff49e74e09dd38 Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.270625 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" event={"ID":"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389","Type":"ContainerStarted","Data":"fb4000f3f794ce5f11cfbb75b7107a420ce00419f0e3e7bc18664f5e0d1889bd"} Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.273467 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ac539301-c457-4332-ad69-87e68d920938","Type":"ContainerStarted","Data":"1d40372c3bcbfe3151f3387ff2efcc37fd27049467de43980eb9e8a7c38c64ea"} Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.277378 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78b87b9446-n4fxp" event={"ID":"421c2c04-702f-403f-a6b0-fa784dd1af3f","Type":"ContainerStarted","Data":"08cca30da0950126763615468a8e5086c415dffac4f3901edc196c07e58e1f86"} Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.277492 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.277526 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.289365 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" event={"ID":"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb","Type":"ContainerDied","Data":"5b023608b1258b6dd6f994df3ed12e940e546c9ff22309d219a290d87d043f4e"} Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.289522 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b023608b1258b6dd6f994df3ed12e940e546c9ff22309d219a290d87d043f4e" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.298904 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" event={"ID":"f721aeee-7daf-413d-885b-5c146881eb99","Type":"ContainerStarted","Data":"7531fc9fa543edd2f37229ea844e5959dcdbc9b72fa510c717ff49e74e09dd38"} Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.310665 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" event={"ID":"8a7aa3be-d2ee-415a-87ac-bb65176dee59","Type":"ContainerStarted","Data":"0f7a20129d84c55ad86f98c7e805d29ffc41c4cb7e52c954861c6e8c1caa15ad"} Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.324165 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-78b87b9446-n4fxp" podStartSLOduration=4.3241368829999995 podStartE2EDuration="4.324136883s" podCreationTimestamp="2026-01-26 09:24:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:24.298695101 +0000 UTC m=+997.607534902" watchObservedRunningTime="2026-01-26 09:24:24.324136883 +0000 UTC m=+997.632976684" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.451226 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.564347 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-config\") pod \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.564944 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bps42\" (UniqueName: \"kubernetes.io/projected/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-kube-api-access-bps42\") pod \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.564995 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-dns-swift-storage-0\") pod \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.565058 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-ovsdbserver-nb\") pod \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.565122 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-dns-svc\") pod \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.565867 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-ovsdbserver-sb\") pod \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\" (UID: \"5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb\") " Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.570996 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-kube-api-access-bps42" (OuterVolumeSpecName: "kube-api-access-bps42") pod "5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb" (UID: "5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb"). InnerVolumeSpecName "kube-api-access-bps42". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.609538 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb" (UID: "5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.610866 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-config" (OuterVolumeSpecName: "config") pod "5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb" (UID: "5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.613592 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb" (UID: "5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.625633 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb" (UID: "5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.645178 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb" (UID: "5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.668677 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.668707 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.668718 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bps42\" (UniqueName: \"kubernetes.io/projected/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-kube-api-access-bps42\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.668727 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.668739 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:24 crc kubenswrapper[4872]: I0126 09:24:24.668749 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.073585 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-conductor-0"] Jan 26 09:24:25 crc kubenswrapper[4872]: W0126 09:24:25.110771 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode135958c_61c2_4199_94b3_2a8e6623310c.slice/crio-85f5d245444bec6fc387176d40ffe579f4526c4253a7602c3ea05e9b069a31f2 WatchSource:0}: Error finding container 85f5d245444bec6fc387176d40ffe579f4526c4253a7602c3ea05e9b069a31f2: Status 404 returned error can't find the container with id 85f5d245444bec6fc387176d40ffe579f4526c4253a7602c3ea05e9b069a31f2 Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.354169 4872 generic.go:334] "Generic (PLEG): container finished" podID="cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2" containerID="34f22c5846eb4b42bbf6fc56fe6e362757e1015f73a401c5bd8a46fcd3cab25b" exitCode=0 Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.355119 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-mhtml" event={"ID":"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2","Type":"ContainerDied","Data":"34f22c5846eb4b42bbf6fc56fe6e362757e1015f73a401c5bd8a46fcd3cab25b"} Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.377467 4872 generic.go:334] "Generic (PLEG): container finished" podID="8a7aa3be-d2ee-415a-87ac-bb65176dee59" containerID="7025f66c80d638b994c0a2de1d945d0fc7e34a9119ce3a2e383de1d4b7dd9285" exitCode=0 Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.377550 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" event={"ID":"8a7aa3be-d2ee-415a-87ac-bb65176dee59","Type":"ContainerDied","Data":"7025f66c80d638b994c0a2de1d945d0fc7e34a9119ce3a2e383de1d4b7dd9285"} Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.401918 4872 generic.go:334] "Generic (PLEG): container finished" podID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerID="02b520f81790fe052d10b40bbca3e30c9b45bb327c1aab132b4362a8bbda7d52" exitCode=0 Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.402036 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e","Type":"ContainerDied","Data":"02b520f81790fe052d10b40bbca3e30c9b45bb327c1aab132b4362a8bbda7d52"} Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.406114 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5f44467695-j8db6" event={"ID":"1dc34b32-f172-4b29-8bd1-a9f7a86d0996","Type":"ContainerStarted","Data":"58767383e88c3cdc455d06a97c6a08c6ac64a82eed2e60af7f322029b7e5b87d"} Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.421949 4872 generic.go:334] "Generic (PLEG): container finished" podID="99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" containerID="b8dafe34e2d01960c568c047f2f3f07cb31abbcbca4a0fd41efbeb28bb54a2de" exitCode=0 Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.422443 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" event={"ID":"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389","Type":"ContainerDied","Data":"b8dafe34e2d01960c568c047f2f3f07cb31abbcbca4a0fd41efbeb28bb54a2de"} Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.425528 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"e135958c-61c2-4199-94b3-2a8e6623310c","Type":"ContainerStarted","Data":"85f5d245444bec6fc387176d40ffe579f4526c4253a7602c3ea05e9b069a31f2"} Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.441637 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bc6dfb648-7gm8x" event={"ID":"2b1a1c03-c8bb-44ab-b607-722109c01b5c","Type":"ContainerStarted","Data":"5a36b336e8d85864c8e9db71848fedf896327122bdf23e72611fd1bef6a14a96"} Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.445570 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5f44467695-j8db6" podStartSLOduration=1.975486452 podStartE2EDuration="5.445541516s" podCreationTimestamp="2026-01-26 09:24:20 +0000 UTC" firstStartedPulling="2026-01-26 09:24:21.079551288 +0000 UTC m=+994.388391089" lastFinishedPulling="2026-01-26 09:24:24.549606352 +0000 UTC m=+997.858446153" observedRunningTime="2026-01-26 09:24:25.432163152 +0000 UTC m=+998.741002953" watchObservedRunningTime="2026-01-26 09:24:25.445541516 +0000 UTC m=+998.754381317" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.481305 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-vbpqn" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.482583 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-664886958b-t96tw" event={"ID":"6532ea83-7260-4883-8cdd-236c90c5c90a","Type":"ContainerStarted","Data":"304c6af015e03f6a75f29ec9b1bd8caec13f1fc1416c12f44a1f69263e87ee49"} Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.483243 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-664886958b-t96tw" event={"ID":"6532ea83-7260-4883-8cdd-236c90c5c90a","Type":"ContainerStarted","Data":"a4bba836cbbf0cee3a27cc53cb59c7948068614a4122768d2db4370f39b77aa6"} Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.525529 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-664886958b-t96tw" podStartSLOduration=2.54211482 podStartE2EDuration="5.525506579s" podCreationTimestamp="2026-01-26 09:24:20 +0000 UTC" firstStartedPulling="2026-01-26 09:24:21.482087494 +0000 UTC m=+994.790927295" lastFinishedPulling="2026-01-26 09:24:24.465479243 +0000 UTC m=+997.774319054" observedRunningTime="2026-01-26 09:24:25.521626359 +0000 UTC m=+998.830466180" watchObservedRunningTime="2026-01-26 09:24:25.525506579 +0000 UTC m=+998.834346370" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.670706 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-vbpqn"] Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.690283 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-vbpqn"] Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.717700 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.803377 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-scripts\") pod \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.807627 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-sg-core-conf-yaml\") pod \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.808999 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-scripts" (OuterVolumeSpecName: "scripts") pod "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" (UID: "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.810061 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-run-httpd\") pod \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.810559 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" (UID: "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.810712 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-log-httpd\") pod \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.811094 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" (UID: "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.811206 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-config-data\") pod \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.811667 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qq65j\" (UniqueName: \"kubernetes.io/projected/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-kube-api-access-qq65j\") pod \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.811704 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-combined-ca-bundle\") pod \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\" (UID: \"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e\") " Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.812763 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.812785 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.812894 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.823078 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-kube-api-access-qq65j" (OuterVolumeSpecName: "kube-api-access-qq65j") pod "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" (UID: "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e"). InnerVolumeSpecName "kube-api-access-qq65j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.857848 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" (UID: "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.914953 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qq65j\" (UniqueName: \"kubernetes.io/projected/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-kube-api-access-qq65j\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.914986 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.943969 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" (UID: "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:25 crc kubenswrapper[4872]: I0126 09:24:25.966350 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-config-data" (OuterVolumeSpecName: "config-data") pod "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" (UID: "25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.016550 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.016701 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.495080 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" event={"ID":"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389","Type":"ContainerStarted","Data":"4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d"} Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.495556 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.496910 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ac539301-c457-4332-ad69-87e68d920938","Type":"ContainerStarted","Data":"ea0887a1e4a593d01a0c3751baf725a437fd548d0eaf0e1992e790f070b19b91"} Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.499176 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"e135958c-61c2-4199-94b3-2a8e6623310c","Type":"ContainerStarted","Data":"4863ce39cb82ea0b2882ea44b9654f7b87f23d069a7df77947df739ae0d506ee"} Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.500396 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e205129d-e73c-4c6f-9110-788265aaada0","Type":"ContainerStarted","Data":"2bdb4897c6dd62d67c69e9b6cbdf3b2bfa248463abe58f6e4321ee3cdf683ff9"} Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.503980 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e","Type":"ContainerDied","Data":"c9e91b0ec85022369e15f303bd12281c20ce037f3a49dd8abf09654957f60857"} Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.504027 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.504096 4872 scope.go:117] "RemoveContainer" containerID="bd643af7e6e14e0c065addf247a3ba8e0f02b3186d10c0d31402c4fa4a6ddb22" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.509323 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5f44467695-j8db6" event={"ID":"1dc34b32-f172-4b29-8bd1-a9f7a86d0996","Type":"ContainerStarted","Data":"15782e64e5280dd7f3c25f5a579723fee2a9018f7c8bca097f8da8837f80825e"} Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.529874 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" podStartSLOduration=5.529844655 podStartE2EDuration="5.529844655s" podCreationTimestamp="2026-01-26 09:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:26.515309682 +0000 UTC m=+999.824149483" watchObservedRunningTime="2026-01-26 09:24:26.529844655 +0000 UTC m=+999.838684456" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.600888 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.626710 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.665574 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:24:26 crc kubenswrapper[4872]: E0126 09:24:26.666048 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="ceilometer-notification-agent" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.666062 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="ceilometer-notification-agent" Jan 26 09:24:26 crc kubenswrapper[4872]: E0126 09:24:26.666075 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="proxy-httpd" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.666080 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="proxy-httpd" Jan 26 09:24:26 crc kubenswrapper[4872]: E0126 09:24:26.666110 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="ceilometer-central-agent" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.666117 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="ceilometer-central-agent" Jan 26 09:24:26 crc kubenswrapper[4872]: E0126 09:24:26.666128 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="sg-core" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.666134 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="sg-core" Jan 26 09:24:26 crc kubenswrapper[4872]: E0126 09:24:26.666153 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb" containerName="init" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.666159 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb" containerName="init" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.666331 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb" containerName="init" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.666350 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="ceilometer-central-agent" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.666361 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="ceilometer-notification-agent" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.666370 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="sg-core" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.666384 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" containerName="proxy-httpd" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.668015 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.685347 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.685846 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.731787 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.743270 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6d0ffab-33e4-4cce-8517-9d081b31f695-log-httpd\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.743339 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-config-data\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.743387 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-scripts\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.743922 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlj7l\" (UniqueName: \"kubernetes.io/projected/f6d0ffab-33e4-4cce-8517-9d081b31f695-kube-api-access-mlj7l\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.744073 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.744250 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.745056 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6d0ffab-33e4-4cce-8517-9d081b31f695-run-httpd\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.847090 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6d0ffab-33e4-4cce-8517-9d081b31f695-run-httpd\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.847168 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6d0ffab-33e4-4cce-8517-9d081b31f695-log-httpd\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.847195 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-config-data\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.847228 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-scripts\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.847262 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlj7l\" (UniqueName: \"kubernetes.io/projected/f6d0ffab-33e4-4cce-8517-9d081b31f695-kube-api-access-mlj7l\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.847297 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.847321 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.849510 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6d0ffab-33e4-4cce-8517-9d081b31f695-run-httpd\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.849867 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6d0ffab-33e4-4cce-8517-9d081b31f695-log-httpd\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.860364 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.860761 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.860882 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-config-data\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.878013 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-scripts\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:26 crc kubenswrapper[4872]: I0126 09:24:26.878082 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlj7l\" (UniqueName: \"kubernetes.io/projected/f6d0ffab-33e4-4cce-8517-9d081b31f695-kube-api-access-mlj7l\") pod \"ceilometer-0\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " pod="openstack/ceilometer-0" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.015526 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.196002 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e" path="/var/lib/kubelet/pods/25e97ec7-d3aa-40f1-b74b-f0fa51cbc75e/volumes" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.197021 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb" path="/var/lib/kubelet/pods/5ab29b5e-e3c0-4419-91fe-fbdbdd411bcb/volumes" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.525590 4872 scope.go:117] "RemoveContainer" containerID="a6d9b0e2833d69a3980ed93e23f6cb81343965d9706ed1f8745a3796fe532e86" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.532724 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" event={"ID":"8a7aa3be-d2ee-415a-87ac-bb65176dee59","Type":"ContainerDied","Data":"0f7a20129d84c55ad86f98c7e805d29ffc41c4cb7e52c954861c6e8c1caa15ad"} Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.532770 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f7a20129d84c55ad86f98c7e805d29ffc41c4cb7e52c954861c6e8c1caa15ad" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.537375 4872 generic.go:334] "Generic (PLEG): container finished" podID="e135958c-61c2-4199-94b3-2a8e6623310c" containerID="4863ce39cb82ea0b2882ea44b9654f7b87f23d069a7df77947df739ae0d506ee" exitCode=0 Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.537504 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"e135958c-61c2-4199-94b3-2a8e6623310c","Type":"ContainerDied","Data":"4863ce39cb82ea0b2882ea44b9654f7b87f23d069a7df77947df739ae0d506ee"} Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.549656 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e205129d-e73c-4c6f-9110-788265aaada0","Type":"ContainerStarted","Data":"ed8fbb207e7669db24c7c13fa73178a64e74c44be185bdf00748c77ff607c235"} Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.585717 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-mhtml" event={"ID":"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2","Type":"ContainerDied","Data":"4a317b5da5825148cbdabe98b670b45db1ce3c594aded7a03da572c242eb7ddb"} Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.585856 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a317b5da5825148cbdabe98b670b45db1ce3c594aded7a03da572c242eb7ddb" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.664344 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.76895825 podStartE2EDuration="6.664326233s" podCreationTimestamp="2026-01-26 09:24:21 +0000 UTC" firstStartedPulling="2026-01-26 09:24:22.656406745 +0000 UTC m=+995.965246546" lastFinishedPulling="2026-01-26 09:24:24.551774728 +0000 UTC m=+997.860614529" observedRunningTime="2026-01-26 09:24:27.657098267 +0000 UTC m=+1000.965938068" watchObservedRunningTime="2026-01-26 09:24:27.664326233 +0000 UTC m=+1000.973166034" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.666664 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.711463 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-mhtml" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.784881 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpsbf\" (UniqueName: \"kubernetes.io/projected/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2-kube-api-access-jpsbf\") pod \"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2\" (UID: \"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2\") " Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.785100 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk96c\" (UniqueName: \"kubernetes.io/projected/8a7aa3be-d2ee-415a-87ac-bb65176dee59-kube-api-access-lk96c\") pod \"8a7aa3be-d2ee-415a-87ac-bb65176dee59\" (UID: \"8a7aa3be-d2ee-415a-87ac-bb65176dee59\") " Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.785205 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2-operator-scripts\") pod \"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2\" (UID: \"cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2\") " Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.785409 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a7aa3be-d2ee-415a-87ac-bb65176dee59-operator-scripts\") pod \"8a7aa3be-d2ee-415a-87ac-bb65176dee59\" (UID: \"8a7aa3be-d2ee-415a-87ac-bb65176dee59\") " Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.786456 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a7aa3be-d2ee-415a-87ac-bb65176dee59-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a7aa3be-d2ee-415a-87ac-bb65176dee59" (UID: "8a7aa3be-d2ee-415a-87ac-bb65176dee59"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.787174 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2" (UID: "cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.807906 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2-kube-api-access-jpsbf" (OuterVolumeSpecName: "kube-api-access-jpsbf") pod "cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2" (UID: "cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2"). InnerVolumeSpecName "kube-api-access-jpsbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.809851 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a7aa3be-d2ee-415a-87ac-bb65176dee59-kube-api-access-lk96c" (OuterVolumeSpecName: "kube-api-access-lk96c") pod "8a7aa3be-d2ee-415a-87ac-bb65176dee59" (UID: "8a7aa3be-d2ee-415a-87ac-bb65176dee59"). InnerVolumeSpecName "kube-api-access-lk96c". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.896220 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk96c\" (UniqueName: \"kubernetes.io/projected/8a7aa3be-d2ee-415a-87ac-bb65176dee59-kube-api-access-lk96c\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.896252 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.896266 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a7aa3be-d2ee-415a-87ac-bb65176dee59-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:27.896279 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpsbf\" (UniqueName: \"kubernetes.io/projected/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2-kube-api-access-jpsbf\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.601064 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-6785557d58-cspfm"] Jan 26 09:24:28 crc kubenswrapper[4872]: E0126 09:24:28.601513 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2" containerName="mariadb-database-create" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.601534 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2" containerName="mariadb-database-create" Jan 26 09:24:28 crc kubenswrapper[4872]: E0126 09:24:28.601548 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a7aa3be-d2ee-415a-87ac-bb65176dee59" containerName="mariadb-account-create-update" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.601557 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a7aa3be-d2ee-415a-87ac-bb65176dee59" containerName="mariadb-account-create-update" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.601739 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2" containerName="mariadb-database-create" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.601759 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a7aa3be-d2ee-415a-87ac-bb65176dee59" containerName="mariadb-account-create-update" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.603080 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.617022 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-b7e6-account-create-update-mdh5l" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.617037 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-mhtml" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.620718 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-internal-svc" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.626309 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-public-svc" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.641809 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-6785557d58-cspfm"] Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.716378 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-public-tls-certs\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.716484 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-config-data\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.716513 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-combined-ca-bundle\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.716570 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30773720-c572-485d-8ea3-3ad3293e4869-logs\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.716595 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-internal-tls-certs\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.716641 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-config-data-custom\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.716699 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/30773720-c572-485d-8ea3-3ad3293e4869-config-data-merged\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.716738 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/30773720-c572-485d-8ea3-3ad3293e4869-etc-podinfo\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.716760 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8jpp\" (UniqueName: \"kubernetes.io/projected/30773720-c572-485d-8ea3-3ad3293e4869-kube-api-access-h8jpp\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.716920 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-scripts\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.777858 4872 scope.go:117] "RemoveContainer" containerID="02b520f81790fe052d10b40bbca3e30c9b45bb327c1aab132b4362a8bbda7d52" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.824589 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30773720-c572-485d-8ea3-3ad3293e4869-logs\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.824696 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-internal-tls-certs\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.824900 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-config-data-custom\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.827036 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/30773720-c572-485d-8ea3-3ad3293e4869-config-data-merged\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.827071 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/30773720-c572-485d-8ea3-3ad3293e4869-etc-podinfo\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.827116 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8jpp\" (UniqueName: \"kubernetes.io/projected/30773720-c572-485d-8ea3-3ad3293e4869-kube-api-access-h8jpp\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.825303 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30773720-c572-485d-8ea3-3ad3293e4869-logs\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.827435 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-scripts\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.827541 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/30773720-c572-485d-8ea3-3ad3293e4869-config-data-merged\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.828230 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-public-tls-certs\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.828385 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-config-data\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.828436 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-combined-ca-bundle\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.832202 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-public-tls-certs\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.836943 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-scripts\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.837566 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-config-data-custom\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.839889 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-internal-tls-certs\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.847251 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-config-data\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.848047 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/30773720-c572-485d-8ea3-3ad3293e4869-etc-podinfo\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.851744 4872 scope.go:117] "RemoveContainer" containerID="6aa5198368a8ddf875c04c2c19c0ee31224b4ec54781dc7d825df99106c52642" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.852865 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8jpp\" (UniqueName: \"kubernetes.io/projected/30773720-c572-485d-8ea3-3ad3293e4869-kube-api-access-h8jpp\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.853844 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30773720-c572-485d-8ea3-3ad3293e4869-combined-ca-bundle\") pod \"ironic-6785557d58-cspfm\" (UID: \"30773720-c572-485d-8ea3-3ad3293e4869\") " pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:28 crc kubenswrapper[4872]: I0126 09:24:28.983306 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:29 crc kubenswrapper[4872]: I0126 09:24:29.282918 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:24:29 crc kubenswrapper[4872]: I0126 09:24:29.657760 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6d0ffab-33e4-4cce-8517-9d081b31f695","Type":"ContainerStarted","Data":"ae7ab075617029cfe3463f3e6cb8e14eeddae87939c4998f8cfb2a396f562e78"} Jan 26 09:24:29 crc kubenswrapper[4872]: I0126 09:24:29.685664 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bc6dfb648-7gm8x" event={"ID":"2b1a1c03-c8bb-44ab-b607-722109c01b5c","Type":"ContainerStarted","Data":"0cfb2006217b2ca01dba2c460e11f24eb4ef47861323bd627c96dc867972de97"} Jan 26 09:24:29 crc kubenswrapper[4872]: I0126 09:24:29.702607 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" event={"ID":"f721aeee-7daf-413d-885b-5c146881eb99","Type":"ContainerStarted","Data":"6f4e993edf18dbf2c209d16d44199aa390f4076942602556b0bd58f3cffb4f1e"} Jan 26 09:24:29 crc kubenswrapper[4872]: I0126 09:24:29.703635 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:29 crc kubenswrapper[4872]: I0126 09:24:29.765097 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-6785557d58-cspfm"] Jan 26 09:24:29 crc kubenswrapper[4872]: I0126 09:24:29.787506 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" podStartSLOduration=4.196973974 podStartE2EDuration="8.787484566s" podCreationTimestamp="2026-01-26 09:24:21 +0000 UTC" firstStartedPulling="2026-01-26 09:24:24.247197328 +0000 UTC m=+997.556037129" lastFinishedPulling="2026-01-26 09:24:28.83770792 +0000 UTC m=+1002.146547721" observedRunningTime="2026-01-26 09:24:29.748982357 +0000 UTC m=+1003.057822168" watchObservedRunningTime="2026-01-26 09:24:29.787484566 +0000 UTC m=+1003.096324367" Jan 26 09:24:30 crc kubenswrapper[4872]: I0126 09:24:30.773181 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ac539301-c457-4332-ad69-87e68d920938","Type":"ContainerStarted","Data":"e9a23a86ce2e839618c41cd3564aea1c010dfa1af9f6e9d6c935a0e338c6df0f"} Jan 26 09:24:30 crc kubenswrapper[4872]: I0126 09:24:30.774288 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ac539301-c457-4332-ad69-87e68d920938" containerName="cinder-api-log" containerID="cri-o://ea0887a1e4a593d01a0c3751baf725a437fd548d0eaf0e1992e790f070b19b91" gracePeriod=30 Jan 26 09:24:30 crc kubenswrapper[4872]: I0126 09:24:30.774379 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 26 09:24:30 crc kubenswrapper[4872]: I0126 09:24:30.774771 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ac539301-c457-4332-ad69-87e68d920938" containerName="cinder-api" containerID="cri-o://e9a23a86ce2e839618c41cd3564aea1c010dfa1af9f6e9d6c935a0e338c6df0f" gracePeriod=30 Jan 26 09:24:30 crc kubenswrapper[4872]: I0126 09:24:30.787072 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-6785557d58-cspfm" event={"ID":"30773720-c572-485d-8ea3-3ad3293e4869","Type":"ContainerStarted","Data":"a845b9ca71fb6bb3eadabe8809b9f03463d468b74a027fe15f1f17a8f3383615"} Jan 26 09:24:30 crc kubenswrapper[4872]: I0126 09:24:30.787134 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-6785557d58-cspfm" event={"ID":"30773720-c572-485d-8ea3-3ad3293e4869","Type":"ContainerStarted","Data":"088efa056c0578199289e437e4b39d7e86a63bbc8bf5a5f7ca64b966757929ae"} Jan 26 09:24:30 crc kubenswrapper[4872]: I0126 09:24:30.810046 4872 generic.go:334] "Generic (PLEG): container finished" podID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerID="0cfb2006217b2ca01dba2c460e11f24eb4ef47861323bd627c96dc867972de97" exitCode=0 Jan 26 09:24:30 crc kubenswrapper[4872]: I0126 09:24:30.811266 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bc6dfb648-7gm8x" event={"ID":"2b1a1c03-c8bb-44ab-b607-722109c01b5c","Type":"ContainerDied","Data":"0cfb2006217b2ca01dba2c460e11f24eb4ef47861323bd627c96dc867972de97"} Jan 26 09:24:30 crc kubenswrapper[4872]: I0126 09:24:30.828642 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6d0ffab-33e4-4cce-8517-9d081b31f695","Type":"ContainerStarted","Data":"155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486"} Jan 26 09:24:30 crc kubenswrapper[4872]: I0126 09:24:30.869863 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=9.869824855 podStartE2EDuration="9.869824855s" podCreationTimestamp="2026-01-26 09:24:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:30.858619887 +0000 UTC m=+1004.167459688" watchObservedRunningTime="2026-01-26 09:24:30.869824855 +0000 UTC m=+1004.178664656" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.264351 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-75b468fcb8-mdk2p"] Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.266355 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-75b468fcb8-mdk2p"] Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.266450 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.277270 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.277554 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.420380 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-config-data-custom\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.420912 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-config-data\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.421030 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-combined-ca-bundle\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.421242 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-internal-tls-certs\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.421566 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-public-tls-certs\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.421655 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0f0f467-e8ad-4766-86e5-b46271a39b39-logs\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.421827 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkhr9\" (UniqueName: \"kubernetes.io/projected/b0f0f467-e8ad-4766-86e5-b46271a39b39-kube-api-access-nkhr9\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.525564 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-combined-ca-bundle\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.525651 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-internal-tls-certs\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.525710 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0f0f467-e8ad-4766-86e5-b46271a39b39-logs\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.525732 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-public-tls-certs\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.525765 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkhr9\" (UniqueName: \"kubernetes.io/projected/b0f0f467-e8ad-4766-86e5-b46271a39b39-kube-api-access-nkhr9\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.525834 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-config-data-custom\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.525935 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-config-data\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.529058 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b0f0f467-e8ad-4766-86e5-b46271a39b39-logs\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.539103 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-config-data-custom\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.540034 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-config-data\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.543947 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-combined-ca-bundle\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.546592 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-internal-tls-certs\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.570025 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0f0f467-e8ad-4766-86e5-b46271a39b39-public-tls-certs\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.571085 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkhr9\" (UniqueName: \"kubernetes.io/projected/b0f0f467-e8ad-4766-86e5-b46271a39b39-kube-api-access-nkhr9\") pod \"barbican-api-75b468fcb8-mdk2p\" (UID: \"b0f0f467-e8ad-4766-86e5-b46271a39b39\") " pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.614895 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.760892 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.862151 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6d0ffab-33e4-4cce-8517-9d081b31f695","Type":"ContainerStarted","Data":"51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54"} Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.879026 4872 generic.go:334] "Generic (PLEG): container finished" podID="ac539301-c457-4332-ad69-87e68d920938" containerID="ea0887a1e4a593d01a0c3751baf725a437fd548d0eaf0e1992e790f070b19b91" exitCode=143 Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.879141 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ac539301-c457-4332-ad69-87e68d920938","Type":"ContainerDied","Data":"ea0887a1e4a593d01a0c3751baf725a437fd548d0eaf0e1992e790f070b19b91"} Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.911632 4872 generic.go:334] "Generic (PLEG): container finished" podID="30773720-c572-485d-8ea3-3ad3293e4869" containerID="a845b9ca71fb6bb3eadabe8809b9f03463d468b74a027fe15f1f17a8f3383615" exitCode=0 Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.911751 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-6785557d58-cspfm" event={"ID":"30773720-c572-485d-8ea3-3ad3293e4869","Type":"ContainerDied","Data":"a845b9ca71fb6bb3eadabe8809b9f03463d468b74a027fe15f1f17a8f3383615"} Jan 26 09:24:31 crc kubenswrapper[4872]: I0126 09:24:31.937745 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bc6dfb648-7gm8x" event={"ID":"2b1a1c03-c8bb-44ab-b607-722109c01b5c","Type":"ContainerStarted","Data":"a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743"} Jan 26 09:24:32 crc kubenswrapper[4872]: I0126 09:24:32.263675 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:32 crc kubenswrapper[4872]: I0126 09:24:32.268308 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 26 09:24:32 crc kubenswrapper[4872]: I0126 09:24:32.283022 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:24:32 crc kubenswrapper[4872]: I0126 09:24:32.380907 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 09:24:32 crc kubenswrapper[4872]: I0126 09:24:32.443724 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-v8rr8"] Jan 26 09:24:32 crc kubenswrapper[4872]: I0126 09:24:32.444032 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" podUID="b62ba569-ac64-47fc-b64f-284d6f51391a" containerName="dnsmasq-dns" containerID="cri-o://c65e3c2aab6a8002de12b441952f0f127e21e28c7b042d72727c0750261d8720" gracePeriod=10 Jan 26 09:24:32 crc kubenswrapper[4872]: I0126 09:24:32.510383 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-75b468fcb8-mdk2p"] Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.007167 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75b468fcb8-mdk2p" event={"ID":"b0f0f467-e8ad-4766-86e5-b46271a39b39","Type":"ContainerStarted","Data":"c7f57c7919f9e5784dba6d0ee3ad98a74ee2298c81196ca10fa06ba6df148dad"} Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.031681 4872 generic.go:334] "Generic (PLEG): container finished" podID="b62ba569-ac64-47fc-b64f-284d6f51391a" containerID="c65e3c2aab6a8002de12b441952f0f127e21e28c7b042d72727c0750261d8720" exitCode=0 Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.031925 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" event={"ID":"b62ba569-ac64-47fc-b64f-284d6f51391a","Type":"ContainerDied","Data":"c65e3c2aab6a8002de12b441952f0f127e21e28c7b042d72727c0750261d8720"} Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.062647 4872 generic.go:334] "Generic (PLEG): container finished" podID="ac539301-c457-4332-ad69-87e68d920938" containerID="e9a23a86ce2e839618c41cd3564aea1c010dfa1af9f6e9d6c935a0e338c6df0f" exitCode=0 Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.062715 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ac539301-c457-4332-ad69-87e68d920938","Type":"ContainerDied","Data":"e9a23a86ce2e839618c41cd3564aea1c010dfa1af9f6e9d6c935a0e338c6df0f"} Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.143894 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-6785557d58-cspfm" event={"ID":"30773720-c572-485d-8ea3-3ad3293e4869","Type":"ContainerStarted","Data":"58697b90f07eb2a1aeda00404a26033e318e96feaab49517f277bb9146f6b2fc"} Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.153186 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.186126 4872 generic.go:334] "Generic (PLEG): container finished" podID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerID="c5c6f2be1c383553efa0b88db15a92ae6889e071a08d69b94b1e6272e2dfb26a" exitCode=1 Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.186230 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bc6dfb648-7gm8x" event={"ID":"2b1a1c03-c8bb-44ab-b607-722109c01b5c","Type":"ContainerDied","Data":"c5c6f2be1c383553efa0b88db15a92ae6889e071a08d69b94b1e6272e2dfb26a"} Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.187124 4872 scope.go:117] "RemoveContainer" containerID="c5c6f2be1c383553efa0b88db15a92ae6889e071a08d69b94b1e6272e2dfb26a" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.191471 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.224038 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e205129d-e73c-4c6f-9110-788265aaada0" containerName="cinder-scheduler" containerID="cri-o://2bdb4897c6dd62d67c69e9b6cbdf3b2bfa248463abe58f6e4321ee3cdf683ff9" gracePeriod=30 Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.225164 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e205129d-e73c-4c6f-9110-788265aaada0" containerName="probe" containerID="cri-o://ed8fbb207e7669db24c7c13fa73178a64e74c44be185bdf00748c77ff607c235" gracePeriod=30 Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.238510 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6d0ffab-33e4-4cce-8517-9d081b31f695","Type":"ContainerStarted","Data":"b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90"} Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.295262 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-combined-ca-bundle\") pod \"ac539301-c457-4332-ad69-87e68d920938\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.295304 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-config-data\") pod \"ac539301-c457-4332-ad69-87e68d920938\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.295335 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-scripts\") pod \"ac539301-c457-4332-ad69-87e68d920938\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.295446 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac539301-c457-4332-ad69-87e68d920938-logs\") pod \"ac539301-c457-4332-ad69-87e68d920938\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.295483 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac539301-c457-4332-ad69-87e68d920938-etc-machine-id\") pod \"ac539301-c457-4332-ad69-87e68d920938\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.295538 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-config-data-custom\") pod \"ac539301-c457-4332-ad69-87e68d920938\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.295615 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pns7\" (UniqueName: \"kubernetes.io/projected/ac539301-c457-4332-ad69-87e68d920938-kube-api-access-8pns7\") pod \"ac539301-c457-4332-ad69-87e68d920938\" (UID: \"ac539301-c457-4332-ad69-87e68d920938\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.299174 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ac539301-c457-4332-ad69-87e68d920938-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ac539301-c457-4332-ad69-87e68d920938" (UID: "ac539301-c457-4332-ad69-87e68d920938"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.315956 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac539301-c457-4332-ad69-87e68d920938-logs" (OuterVolumeSpecName: "logs") pod "ac539301-c457-4332-ad69-87e68d920938" (UID: "ac539301-c457-4332-ad69-87e68d920938"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.330616 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ac539301-c457-4332-ad69-87e68d920938" (UID: "ac539301-c457-4332-ad69-87e68d920938"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.344086 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac539301-c457-4332-ad69-87e68d920938-kube-api-access-8pns7" (OuterVolumeSpecName: "kube-api-access-8pns7") pod "ac539301-c457-4332-ad69-87e68d920938" (UID: "ac539301-c457-4332-ad69-87e68d920938"). InnerVolumeSpecName "kube-api-access-8pns7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.352945 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-scripts" (OuterVolumeSpecName: "scripts") pod "ac539301-c457-4332-ad69-87e68d920938" (UID: "ac539301-c457-4332-ad69-87e68d920938"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.408039 4872 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.408079 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pns7\" (UniqueName: \"kubernetes.io/projected/ac539301-c457-4332-ad69-87e68d920938-kube-api-access-8pns7\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.408094 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.408106 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac539301-c457-4332-ad69-87e68d920938-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.408115 4872 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ac539301-c457-4332-ad69-87e68d920938-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.412146 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac539301-c457-4332-ad69-87e68d920938" (UID: "ac539301-c457-4332-ad69-87e68d920938"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.445039 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-config-data" (OuterVolumeSpecName: "config-data") pod "ac539301-c457-4332-ad69-87e68d920938" (UID: "ac539301-c457-4332-ad69-87e68d920938"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.465024 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-5c4cd99bf6-6gmxf" podUID="5c05094d-a3b9-4016-88c1-2eabf12f3920" containerName="neutron-httpd" probeResult="failure" output="Get \"http://10.217.0.154:9696/\": dial tcp 10.217.0.154:9696: connect: connection refused" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.485308 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.493224 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.512149 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.512209 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac539301-c457-4332-ad69-87e68d920938-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.614508 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-dns-svc\") pod \"b62ba569-ac64-47fc-b64f-284d6f51391a\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.614714 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-ovsdbserver-nb\") pod \"b62ba569-ac64-47fc-b64f-284d6f51391a\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.614789 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-dns-swift-storage-0\") pod \"b62ba569-ac64-47fc-b64f-284d6f51391a\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.615567 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bfpv\" (UniqueName: \"kubernetes.io/projected/b62ba569-ac64-47fc-b64f-284d6f51391a-kube-api-access-7bfpv\") pod \"b62ba569-ac64-47fc-b64f-284d6f51391a\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.615647 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-config\") pod \"b62ba569-ac64-47fc-b64f-284d6f51391a\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.615699 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-ovsdbserver-sb\") pod \"b62ba569-ac64-47fc-b64f-284d6f51391a\" (UID: \"b62ba569-ac64-47fc-b64f-284d6f51391a\") " Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.644125 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b62ba569-ac64-47fc-b64f-284d6f51391a-kube-api-access-7bfpv" (OuterVolumeSpecName: "kube-api-access-7bfpv") pod "b62ba569-ac64-47fc-b64f-284d6f51391a" (UID: "b62ba569-ac64-47fc-b64f-284d6f51391a"). InnerVolumeSpecName "kube-api-access-7bfpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.721648 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bfpv\" (UniqueName: \"kubernetes.io/projected/b62ba569-ac64-47fc-b64f-284d6f51391a-kube-api-access-7bfpv\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.738239 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-config" (OuterVolumeSpecName: "config") pod "b62ba569-ac64-47fc-b64f-284d6f51391a" (UID: "b62ba569-ac64-47fc-b64f-284d6f51391a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.747871 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b62ba569-ac64-47fc-b64f-284d6f51391a" (UID: "b62ba569-ac64-47fc-b64f-284d6f51391a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.813512 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b62ba569-ac64-47fc-b64f-284d6f51391a" (UID: "b62ba569-ac64-47fc-b64f-284d6f51391a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.816496 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b62ba569-ac64-47fc-b64f-284d6f51391a" (UID: "b62ba569-ac64-47fc-b64f-284d6f51391a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.826904 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.828347 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.828462 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.828527 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.832020 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b62ba569-ac64-47fc-b64f-284d6f51391a" (UID: "b62ba569-ac64-47fc-b64f-284d6f51391a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:24:33 crc kubenswrapper[4872]: I0126 09:24:33.939562 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b62ba569-ac64-47fc-b64f-284d6f51391a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.063640 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.320455 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bc6dfb648-7gm8x" event={"ID":"2b1a1c03-c8bb-44ab-b607-722109c01b5c","Type":"ContainerStarted","Data":"726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b"} Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.321988 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:34 crc kubenswrapper[4872]: E0126 09:24:34.352237 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode205129d_e73c_4c6f_9110_788265aaada0.slice/crio-conmon-ed8fbb207e7669db24c7c13fa73178a64e74c44be185bdf00748c77ff607c235.scope\": RecentStats: unable to find data in memory cache]" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.356287 4872 generic.go:334] "Generic (PLEG): container finished" podID="e205129d-e73c-4c6f-9110-788265aaada0" containerID="ed8fbb207e7669db24c7c13fa73178a64e74c44be185bdf00748c77ff607c235" exitCode=0 Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.356430 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e205129d-e73c-4c6f-9110-788265aaada0","Type":"ContainerDied","Data":"ed8fbb207e7669db24c7c13fa73178a64e74c44be185bdf00748c77ff607c235"} Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.361470 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-bc6dfb648-7gm8x" podStartSLOduration=8.823424419 podStartE2EDuration="13.361441283s" podCreationTimestamp="2026-01-26 09:24:21 +0000 UTC" firstStartedPulling="2026-01-26 09:24:24.299169762 +0000 UTC m=+997.608009563" lastFinishedPulling="2026-01-26 09:24:28.837186626 +0000 UTC m=+1002.146026427" observedRunningTime="2026-01-26 09:24:34.346990851 +0000 UTC m=+1007.655830662" watchObservedRunningTime="2026-01-26 09:24:34.361441283 +0000 UTC m=+1007.670281084" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.379391 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75b468fcb8-mdk2p" event={"ID":"b0f0f467-e8ad-4766-86e5-b46271a39b39","Type":"ContainerStarted","Data":"9128237f898346d29a648a9690c8ee5a7fd005caf4029d5db1aeeb6ba36b1e0f"} Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.379439 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-75b468fcb8-mdk2p" event={"ID":"b0f0f467-e8ad-4766-86e5-b46271a39b39","Type":"ContainerStarted","Data":"552f118abc0c5b85aa2d26f7548efd24cd5460fbe0a7876bf5497551260ad3bb"} Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.379482 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.379503 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.389136 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" event={"ID":"b62ba569-ac64-47fc-b64f-284d6f51391a","Type":"ContainerDied","Data":"536bef46f41cd5c4db426998f3b3f486cdc3a87b2fc44bdbf9cf6aa28584a7e5"} Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.389202 4872 scope.go:117] "RemoveContainer" containerID="c65e3c2aab6a8002de12b441952f0f127e21e28c7b042d72727c0750261d8720" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.389370 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.443450 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-75b468fcb8-mdk2p" podStartSLOduration=3.443419297 podStartE2EDuration="3.443419297s" podCreationTimestamp="2026-01-26 09:24:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:34.402372473 +0000 UTC m=+1007.711212274" watchObservedRunningTime="2026-01-26 09:24:34.443419297 +0000 UTC m=+1007.752259098" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.458238 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ac539301-c457-4332-ad69-87e68d920938","Type":"ContainerDied","Data":"1d40372c3bcbfe3151f3387ff2efcc37fd27049467de43980eb9e8a7c38c64ea"} Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.458404 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.481642 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-v8rr8"] Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.487942 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-6785557d58-cspfm" event={"ID":"30773720-c572-485d-8ea3-3ad3293e4869","Type":"ContainerStarted","Data":"786f5964f1965491554abc856b2137b9c640b4bafb3e323bdc20c94af76cdbbc"} Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.488000 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-v8rr8"] Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.488033 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.519985 4872 scope.go:117] "RemoveContainer" containerID="57676e8b2fa37390789272cf322a5505c55467f945a44aca48d0a7be04f22338" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.554021 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-6785557d58-cspfm" podStartSLOduration=6.5539882259999995 podStartE2EDuration="6.553988226s" podCreationTimestamp="2026-01-26 09:24:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:34.528528062 +0000 UTC m=+1007.837367863" watchObservedRunningTime="2026-01-26 09:24:34.553988226 +0000 UTC m=+1007.862828027" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.566550 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.598871 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.606345 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Jan 26 09:24:34 crc kubenswrapper[4872]: E0126 09:24:34.606960 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62ba569-ac64-47fc-b64f-284d6f51391a" containerName="dnsmasq-dns" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.606989 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62ba569-ac64-47fc-b64f-284d6f51391a" containerName="dnsmasq-dns" Jan 26 09:24:34 crc kubenswrapper[4872]: E0126 09:24:34.607022 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac539301-c457-4332-ad69-87e68d920938" containerName="cinder-api-log" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.607031 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac539301-c457-4332-ad69-87e68d920938" containerName="cinder-api-log" Jan 26 09:24:34 crc kubenswrapper[4872]: E0126 09:24:34.607050 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b62ba569-ac64-47fc-b64f-284d6f51391a" containerName="init" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.607057 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="b62ba569-ac64-47fc-b64f-284d6f51391a" containerName="init" Jan 26 09:24:34 crc kubenswrapper[4872]: E0126 09:24:34.607083 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac539301-c457-4332-ad69-87e68d920938" containerName="cinder-api" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.607092 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac539301-c457-4332-ad69-87e68d920938" containerName="cinder-api" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.607541 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac539301-c457-4332-ad69-87e68d920938" containerName="cinder-api" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.607565 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="b62ba569-ac64-47fc-b64f-284d6f51391a" containerName="dnsmasq-dns" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.607583 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac539301-c457-4332-ad69-87e68d920938" containerName="cinder-api-log" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.608779 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.612594 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.612899 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.613016 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.628107 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.665292 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.667998 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-b65567844-p9ckk" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.682700 4872 scope.go:117] "RemoveContainer" containerID="e9a23a86ce2e839618c41cd3564aea1c010dfa1af9f6e9d6c935a0e338c6df0f" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.766927 4872 scope.go:117] "RemoveContainer" containerID="ea0887a1e4a593d01a0c3751baf725a437fd548d0eaf0e1992e790f070b19b91" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.773019 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-scripts\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.773103 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9c67aaa-15db-4359-b978-685a7ed15ec4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.773125 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c67aaa-15db-4359-b978-685a7ed15ec4-logs\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.773155 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.773200 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.773249 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-config-data\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.773266 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-config-data-custom\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.773284 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.773306 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfrpt\" (UniqueName: \"kubernetes.io/projected/d9c67aaa-15db-4359-b978-685a7ed15ec4-kube-api-access-kfrpt\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.876021 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-scripts\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.876121 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9c67aaa-15db-4359-b978-685a7ed15ec4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.876161 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c67aaa-15db-4359-b978-685a7ed15ec4-logs\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.876206 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.876600 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.876662 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-config-data\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.876694 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-config-data-custom\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.876718 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.876749 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfrpt\" (UniqueName: \"kubernetes.io/projected/d9c67aaa-15db-4359-b978-685a7ed15ec4-kube-api-access-kfrpt\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.877546 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9c67aaa-15db-4359-b978-685a7ed15ec4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.878098 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d9c67aaa-15db-4359-b978-685a7ed15ec4-logs\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.897247 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-config-data\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.898385 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-scripts\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.898873 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-config-data-custom\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.902604 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.916936 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.928554 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9c67aaa-15db-4359-b978-685a7ed15ec4-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:34 crc kubenswrapper[4872]: I0126 09:24:34.935540 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfrpt\" (UniqueName: \"kubernetes.io/projected/d9c67aaa-15db-4359-b978-685a7ed15ec4-kube-api-access-kfrpt\") pod \"cinder-api-0\" (UID: \"d9c67aaa-15db-4359-b978-685a7ed15ec4\") " pod="openstack/cinder-api-0" Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.011298 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.201970 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac539301-c457-4332-ad69-87e68d920938" path="/var/lib/kubelet/pods/ac539301-c457-4332-ad69-87e68d920938/volumes" Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.202782 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b62ba569-ac64-47fc-b64f-284d6f51391a" path="/var/lib/kubelet/pods/b62ba569-ac64-47fc-b64f-284d6f51391a/volumes" Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.539511 4872 generic.go:334] "Generic (PLEG): container finished" podID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerID="726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b" exitCode=1 Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.539957 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bc6dfb648-7gm8x" event={"ID":"2b1a1c03-c8bb-44ab-b607-722109c01b5c","Type":"ContainerDied","Data":"726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b"} Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.539998 4872 scope.go:117] "RemoveContainer" containerID="c5c6f2be1c383553efa0b88db15a92ae6889e071a08d69b94b1e6272e2dfb26a" Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.541718 4872 scope.go:117] "RemoveContainer" containerID="726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b" Jan 26 09:24:35 crc kubenswrapper[4872]: E0126 09:24:35.543019 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-api pod=ironic-bc6dfb648-7gm8x_openstack(2b1a1c03-c8bb-44ab-b607-722109c01b5c)\"" pod="openstack/ironic-bc6dfb648-7gm8x" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.555521 4872 generic.go:334] "Generic (PLEG): container finished" podID="f721aeee-7daf-413d-885b-5c146881eb99" containerID="6f4e993edf18dbf2c209d16d44199aa390f4076942602556b0bd58f3cffb4f1e" exitCode=1 Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.555754 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" event={"ID":"f721aeee-7daf-413d-885b-5c146881eb99","Type":"ContainerDied","Data":"6f4e993edf18dbf2c209d16d44199aa390f4076942602556b0bd58f3cffb4f1e"} Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.557065 4872 scope.go:117] "RemoveContainer" containerID="6f4e993edf18dbf2c209d16d44199aa390f4076942602556b0bd58f3cffb4f1e" Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.572905 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6d0ffab-33e4-4cce-8517-9d081b31f695","Type":"ContainerStarted","Data":"6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46"} Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.574673 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.626139 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.890137816 podStartE2EDuration="9.626107243s" podCreationTimestamp="2026-01-26 09:24:26 +0000 UTC" firstStartedPulling="2026-01-26 09:24:29.294161769 +0000 UTC m=+1002.603001570" lastFinishedPulling="2026-01-26 09:24:34.030131196 +0000 UTC m=+1007.338970997" observedRunningTime="2026-01-26 09:24:35.613869368 +0000 UTC m=+1008.922709169" watchObservedRunningTime="2026-01-26 09:24:35.626107243 +0000 UTC m=+1008.934947044" Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.666361 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.852312 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-795668bf-dk6bd" Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.930294 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-88948b444-cjfbr"] Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.931877 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-88948b444-cjfbr" podUID="ab2bb608-c691-45ba-a23b-b0df661b68b7" containerName="neutron-httpd" containerID="cri-o://413e857e7cc2d427c4e6329b01cc361aec394e2fab0a6588eed164991c76e0f0" gracePeriod=30 Jan 26 09:24:35 crc kubenswrapper[4872]: I0126 09:24:35.932177 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-88948b444-cjfbr" podUID="ab2bb608-c691-45ba-a23b-b0df661b68b7" containerName="neutron-api" containerID="cri-o://ef2ebf6d3a564ff492b2caf68378c294a29edd9965cc00af1dc8b28b36f98d7d" gracePeriod=30 Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.316141 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-db-sync-dv876"] Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.318439 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.356987 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.357399 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-sync-dv876"] Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.357395 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.422246 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-combined-ca-bundle\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.422811 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/909a12a7-4b22-43b8-8d96-1fff1ef2738c-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.422937 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ntc4\" (UniqueName: \"kubernetes.io/projected/909a12a7-4b22-43b8-8d96-1fff1ef2738c-kube-api-access-8ntc4\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.423030 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/909a12a7-4b22-43b8-8d96-1fff1ef2738c-var-lib-ironic\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.423115 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/909a12a7-4b22-43b8-8d96-1fff1ef2738c-etc-podinfo\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.423230 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-scripts\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.423329 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-config\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.518744 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-769b85d4b6-8zpg4" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.527186 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/909a12a7-4b22-43b8-8d96-1fff1ef2738c-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.527266 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ntc4\" (UniqueName: \"kubernetes.io/projected/909a12a7-4b22-43b8-8d96-1fff1ef2738c-kube-api-access-8ntc4\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.527296 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/909a12a7-4b22-43b8-8d96-1fff1ef2738c-var-lib-ironic\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.527314 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/909a12a7-4b22-43b8-8d96-1fff1ef2738c-etc-podinfo\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.527338 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-scripts\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.527357 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-config\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.527514 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-combined-ca-bundle\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.528129 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/909a12a7-4b22-43b8-8d96-1fff1ef2738c-var-lib-ironic\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.528219 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/909a12a7-4b22-43b8-8d96-1fff1ef2738c-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.534730 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-scripts\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.534992 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-combined-ca-bundle\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.537272 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-config\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.559295 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/909a12a7-4b22-43b8-8d96-1fff1ef2738c-etc-podinfo\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.575586 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ntc4\" (UniqueName: \"kubernetes.io/projected/909a12a7-4b22-43b8-8d96-1fff1ef2738c-kube-api-access-8ntc4\") pod \"ironic-inspector-db-sync-dv876\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.605168 4872 generic.go:334] "Generic (PLEG): container finished" podID="ab2bb608-c691-45ba-a23b-b0df661b68b7" containerID="413e857e7cc2d427c4e6329b01cc361aec394e2fab0a6588eed164991c76e0f0" exitCode=0 Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.605882 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-88948b444-cjfbr" event={"ID":"ab2bb608-c691-45ba-a23b-b0df661b68b7","Type":"ContainerDied","Data":"413e857e7cc2d427c4e6329b01cc361aec394e2fab0a6588eed164991c76e0f0"} Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.606143 4872 scope.go:117] "RemoveContainer" containerID="726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b" Jan 26 09:24:36 crc kubenswrapper[4872]: E0126 09:24:36.607161 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-api pod=ironic-bc6dfb648-7gm8x_openstack(2b1a1c03-c8bb-44ab-b607-722109c01b5c)\"" pod="openstack/ironic-bc6dfb648-7gm8x" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" Jan 26 09:24:36 crc kubenswrapper[4872]: I0126 09:24:36.660361 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:37 crc kubenswrapper[4872]: I0126 09:24:37.074254 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:37 crc kubenswrapper[4872]: I0126 09:24:37.074691 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:37 crc kubenswrapper[4872]: I0126 09:24:37.311043 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:37 crc kubenswrapper[4872]: I0126 09:24:37.621976 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c4cd99bf6-6gmxf_5c05094d-a3b9-4016-88c1-2eabf12f3920/neutron-api/0.log" Jan 26 09:24:37 crc kubenswrapper[4872]: I0126 09:24:37.622027 4872 generic.go:334] "Generic (PLEG): container finished" podID="5c05094d-a3b9-4016-88c1-2eabf12f3920" containerID="269f313a32b515484c32ad8ae94fc4198626212900df162c04dcd127a11599fd" exitCode=137 Jan 26 09:24:37 crc kubenswrapper[4872]: I0126 09:24:37.622088 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c4cd99bf6-6gmxf" event={"ID":"5c05094d-a3b9-4016-88c1-2eabf12f3920","Type":"ContainerDied","Data":"269f313a32b515484c32ad8ae94fc4198626212900df162c04dcd127a11599fd"} Jan 26 09:24:37 crc kubenswrapper[4872]: I0126 09:24:37.625673 4872 generic.go:334] "Generic (PLEG): container finished" podID="e205129d-e73c-4c6f-9110-788265aaada0" containerID="2bdb4897c6dd62d67c69e9b6cbdf3b2bfa248463abe58f6e4321ee3cdf683ff9" exitCode=0 Jan 26 09:24:37 crc kubenswrapper[4872]: I0126 09:24:37.625760 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e205129d-e73c-4c6f-9110-788265aaada0","Type":"ContainerDied","Data":"2bdb4897c6dd62d67c69e9b6cbdf3b2bfa248463abe58f6e4321ee3cdf683ff9"} Jan 26 09:24:37 crc kubenswrapper[4872]: I0126 09:24:37.627066 4872 scope.go:117] "RemoveContainer" containerID="726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b" Jan 26 09:24:37 crc kubenswrapper[4872]: E0126 09:24:37.627416 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-api pod=ironic-bc6dfb648-7gm8x_openstack(2b1a1c03-c8bb-44ab-b607-722109c01b5c)\"" pod="openstack/ironic-bc6dfb648-7gm8x" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.126349 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-84b966f6c9-v8rr8" podUID="b62ba569-ac64-47fc-b64f-284d6f51391a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.153:5353: i/o timeout" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.602890 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.604967 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.612939 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.613167 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-zdhrv" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.613293 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.644941 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.674882 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.688007 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d9c67aaa-15db-4359-b978-685a7ed15ec4","Type":"ContainerStarted","Data":"79f6562b4f8fe0501e35f69a156023603fba99e7d6372b58938d7cad5618836d"} Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.697417 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.697757 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e205129d-e73c-4c6f-9110-788265aaada0","Type":"ContainerDied","Data":"db4cca304f5ed0fa693cbcadd873d0e0eefe9b9f0690ef0bcb5befbe168d4bfa"} Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.698225 4872 scope.go:117] "RemoveContainer" containerID="ed8fbb207e7669db24c7c13fa73178a64e74c44be185bdf00748c77ff607c235" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.713314 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52a92868-e230-4403-9297-51c2b457343b-openstack-config-secret\") pod \"openstackclient\" (UID: \"52a92868-e230-4403-9297-51c2b457343b\") " pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.713419 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a92868-e230-4403-9297-51c2b457343b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"52a92868-e230-4403-9297-51c2b457343b\") " pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.714769 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52a92868-e230-4403-9297-51c2b457343b-openstack-config\") pod \"openstackclient\" (UID: \"52a92868-e230-4403-9297-51c2b457343b\") " pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.717180 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwncl\" (UniqueName: \"kubernetes.io/projected/52a92868-e230-4403-9297-51c2b457343b-kube-api-access-qwncl\") pod \"openstackclient\" (UID: \"52a92868-e230-4403-9297-51c2b457343b\") " pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.717928 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" event={"ID":"f721aeee-7daf-413d-885b-5c146881eb99","Type":"ContainerStarted","Data":"4e2e1a239f503efca62c57a52ec0ca4f72685e7934b746cee1b12681d3e89b09"} Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.718242 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.745146 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c4cd99bf6-6gmxf_5c05094d-a3b9-4016-88c1-2eabf12f3920/neutron-api/0.log" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.745265 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.812833 4872 scope.go:117] "RemoveContainer" containerID="2bdb4897c6dd62d67c69e9b6cbdf3b2bfa248463abe58f6e4321ee3cdf683ff9" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.818551 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e205129d-e73c-4c6f-9110-788265aaada0-etc-machine-id\") pod \"e205129d-e73c-4c6f-9110-788265aaada0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.818701 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-scripts\") pod \"e205129d-e73c-4c6f-9110-788265aaada0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.818730 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-config-data-custom\") pod \"e205129d-e73c-4c6f-9110-788265aaada0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.818765 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcn24\" (UniqueName: \"kubernetes.io/projected/e205129d-e73c-4c6f-9110-788265aaada0-kube-api-access-fcn24\") pod \"e205129d-e73c-4c6f-9110-788265aaada0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.818836 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-combined-ca-bundle\") pod \"e205129d-e73c-4c6f-9110-788265aaada0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.818870 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-config-data\") pod \"e205129d-e73c-4c6f-9110-788265aaada0\" (UID: \"e205129d-e73c-4c6f-9110-788265aaada0\") " Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.819181 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwncl\" (UniqueName: \"kubernetes.io/projected/52a92868-e230-4403-9297-51c2b457343b-kube-api-access-qwncl\") pod \"openstackclient\" (UID: \"52a92868-e230-4403-9297-51c2b457343b\") " pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.819210 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52a92868-e230-4403-9297-51c2b457343b-openstack-config-secret\") pod \"openstackclient\" (UID: \"52a92868-e230-4403-9297-51c2b457343b\") " pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.819226 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a92868-e230-4403-9297-51c2b457343b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"52a92868-e230-4403-9297-51c2b457343b\") " pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.819253 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52a92868-e230-4403-9297-51c2b457343b-openstack-config\") pod \"openstackclient\" (UID: \"52a92868-e230-4403-9297-51c2b457343b\") " pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.820123 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e205129d-e73c-4c6f-9110-788265aaada0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e205129d-e73c-4c6f-9110-788265aaada0" (UID: "e205129d-e73c-4c6f-9110-788265aaada0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.822673 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/52a92868-e230-4403-9297-51c2b457343b-openstack-config\") pod \"openstackclient\" (UID: \"52a92868-e230-4403-9297-51c2b457343b\") " pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.827420 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/52a92868-e230-4403-9297-51c2b457343b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"52a92868-e230-4403-9297-51c2b457343b\") " pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.827565 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e205129d-e73c-4c6f-9110-788265aaada0" (UID: "e205129d-e73c-4c6f-9110-788265aaada0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.830284 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/52a92868-e230-4403-9297-51c2b457343b-openstack-config-secret\") pod \"openstackclient\" (UID: \"52a92868-e230-4403-9297-51c2b457343b\") " pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.830349 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e205129d-e73c-4c6f-9110-788265aaada0-kube-api-access-fcn24" (OuterVolumeSpecName: "kube-api-access-fcn24") pod "e205129d-e73c-4c6f-9110-788265aaada0" (UID: "e205129d-e73c-4c6f-9110-788265aaada0"). InnerVolumeSpecName "kube-api-access-fcn24". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.844578 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-scripts" (OuterVolumeSpecName: "scripts") pod "e205129d-e73c-4c6f-9110-788265aaada0" (UID: "e205129d-e73c-4c6f-9110-788265aaada0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.845205 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwncl\" (UniqueName: \"kubernetes.io/projected/52a92868-e230-4403-9297-51c2b457343b-kube-api-access-qwncl\") pod \"openstackclient\" (UID: \"52a92868-e230-4403-9297-51c2b457343b\") " pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.920524 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-httpd-config\") pod \"5c05094d-a3b9-4016-88c1-2eabf12f3920\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.920571 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-config\") pod \"5c05094d-a3b9-4016-88c1-2eabf12f3920\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.920632 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-ovndb-tls-certs\") pod \"5c05094d-a3b9-4016-88c1-2eabf12f3920\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.920657 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-combined-ca-bundle\") pod \"5c05094d-a3b9-4016-88c1-2eabf12f3920\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.920855 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llsh5\" (UniqueName: \"kubernetes.io/projected/5c05094d-a3b9-4016-88c1-2eabf12f3920-kube-api-access-llsh5\") pod \"5c05094d-a3b9-4016-88c1-2eabf12f3920\" (UID: \"5c05094d-a3b9-4016-88c1-2eabf12f3920\") " Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.922751 4872 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e205129d-e73c-4c6f-9110-788265aaada0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.922777 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.922786 4872 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.922820 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcn24\" (UniqueName: \"kubernetes.io/projected/e205129d-e73c-4c6f-9110-788265aaada0-kube-api-access-fcn24\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.927174 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "5c05094d-a3b9-4016-88c1-2eabf12f3920" (UID: "5c05094d-a3b9-4016-88c1-2eabf12f3920"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.928481 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-config-data" (OuterVolumeSpecName: "config-data") pod "e205129d-e73c-4c6f-9110-788265aaada0" (UID: "e205129d-e73c-4c6f-9110-788265aaada0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.928975 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c05094d-a3b9-4016-88c1-2eabf12f3920-kube-api-access-llsh5" (OuterVolumeSpecName: "kube-api-access-llsh5") pod "5c05094d-a3b9-4016-88c1-2eabf12f3920" (UID: "5c05094d-a3b9-4016-88c1-2eabf12f3920"). InnerVolumeSpecName "kube-api-access-llsh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.932487 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e205129d-e73c-4c6f-9110-788265aaada0" (UID: "e205129d-e73c-4c6f-9110-788265aaada0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.977699 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c05094d-a3b9-4016-88c1-2eabf12f3920" (UID: "5c05094d-a3b9-4016-88c1-2eabf12f3920"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.985161 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Jan 26 09:24:38 crc kubenswrapper[4872]: I0126 09:24:38.993102 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-config" (OuterVolumeSpecName: "config") pod "5c05094d-a3b9-4016-88c1-2eabf12f3920" (UID: "5c05094d-a3b9-4016-88c1-2eabf12f3920"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.031265 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.031300 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.031309 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e205129d-e73c-4c6f-9110-788265aaada0-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.031318 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llsh5\" (UniqueName: \"kubernetes.io/projected/5c05094d-a3b9-4016-88c1-2eabf12f3920-kube-api-access-llsh5\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.031328 4872 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-httpd-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.031338 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.046778 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-sync-dv876"] Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.086643 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.117858 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "5c05094d-a3b9-4016-88c1-2eabf12f3920" (UID: "5c05094d-a3b9-4016-88c1-2eabf12f3920"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.138817 4872 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c05094d-a3b9-4016-88c1-2eabf12f3920-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.153714 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.247127 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e205129d-e73c-4c6f-9110-788265aaada0" path="/var/lib/kubelet/pods/e205129d-e73c-4c6f-9110-788265aaada0/volumes" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.248426 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 09:24:39 crc kubenswrapper[4872]: E0126 09:24:39.248733 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e205129d-e73c-4c6f-9110-788265aaada0" containerName="cinder-scheduler" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.248751 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e205129d-e73c-4c6f-9110-788265aaada0" containerName="cinder-scheduler" Jan 26 09:24:39 crc kubenswrapper[4872]: E0126 09:24:39.248770 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c05094d-a3b9-4016-88c1-2eabf12f3920" containerName="neutron-api" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.248776 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c05094d-a3b9-4016-88c1-2eabf12f3920" containerName="neutron-api" Jan 26 09:24:39 crc kubenswrapper[4872]: E0126 09:24:39.248788 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e205129d-e73c-4c6f-9110-788265aaada0" containerName="probe" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.248807 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e205129d-e73c-4c6f-9110-788265aaada0" containerName="probe" Jan 26 09:24:39 crc kubenswrapper[4872]: E0126 09:24:39.248840 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c05094d-a3b9-4016-88c1-2eabf12f3920" containerName="neutron-httpd" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.248846 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c05094d-a3b9-4016-88c1-2eabf12f3920" containerName="neutron-httpd" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.249016 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e205129d-e73c-4c6f-9110-788265aaada0" containerName="cinder-scheduler" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.249032 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c05094d-a3b9-4016-88c1-2eabf12f3920" containerName="neutron-httpd" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.249051 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c05094d-a3b9-4016-88c1-2eabf12f3920" containerName="neutron-api" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.249061 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e205129d-e73c-4c6f-9110-788265aaada0" containerName="probe" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.250172 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.250272 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.254393 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.344341 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6faf686b-43c3-44f2-9e84-a142ce8192d1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.344413 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6faf686b-43c3-44f2-9e84-a142ce8192d1-scripts\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.344516 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6faf686b-43c3-44f2-9e84-a142ce8192d1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.344540 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6faf686b-43c3-44f2-9e84-a142ce8192d1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.344561 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6faf686b-43c3-44f2-9e84-a142ce8192d1-config-data\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.344603 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z78s\" (UniqueName: \"kubernetes.io/projected/6faf686b-43c3-44f2-9e84-a142ce8192d1-kube-api-access-4z78s\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.446595 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6faf686b-43c3-44f2-9e84-a142ce8192d1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.446657 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6faf686b-43c3-44f2-9e84-a142ce8192d1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.446686 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6faf686b-43c3-44f2-9e84-a142ce8192d1-config-data\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.446746 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z78s\" (UniqueName: \"kubernetes.io/projected/6faf686b-43c3-44f2-9e84-a142ce8192d1-kube-api-access-4z78s\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.446820 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6faf686b-43c3-44f2-9e84-a142ce8192d1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.446861 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6faf686b-43c3-44f2-9e84-a142ce8192d1-scripts\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.448932 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6faf686b-43c3-44f2-9e84-a142ce8192d1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.475011 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6faf686b-43c3-44f2-9e84-a142ce8192d1-config-data\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.476313 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6faf686b-43c3-44f2-9e84-a142ce8192d1-scripts\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.476777 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6faf686b-43c3-44f2-9e84-a142ce8192d1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.479247 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6faf686b-43c3-44f2-9e84-a142ce8192d1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.484384 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z78s\" (UniqueName: \"kubernetes.io/projected/6faf686b-43c3-44f2-9e84-a142ce8192d1-kube-api-access-4z78s\") pod \"cinder-scheduler-0\" (UID: \"6faf686b-43c3-44f2-9e84-a142ce8192d1\") " pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.665006 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.756952 4872 generic.go:334] "Generic (PLEG): container finished" podID="ab2bb608-c691-45ba-a23b-b0df661b68b7" containerID="ef2ebf6d3a564ff492b2caf68378c294a29edd9965cc00af1dc8b28b36f98d7d" exitCode=0 Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.757324 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-88948b444-cjfbr" event={"ID":"ab2bb608-c691-45ba-a23b-b0df661b68b7","Type":"ContainerDied","Data":"ef2ebf6d3a564ff492b2caf68378c294a29edd9965cc00af1dc8b28b36f98d7d"} Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.761576 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-dv876" event={"ID":"909a12a7-4b22-43b8-8d96-1fff1ef2738c","Type":"ContainerStarted","Data":"adc796bea19110194b6372ef1b76d1f821128af90d2b05fc61b80a2bc3866580"} Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.764615 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5c4cd99bf6-6gmxf_5c05094d-a3b9-4016-88c1-2eabf12f3920/neutron-api/0.log" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.764718 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c4cd99bf6-6gmxf" event={"ID":"5c05094d-a3b9-4016-88c1-2eabf12f3920","Type":"ContainerDied","Data":"a5042937a3f42c82008c2c6a0ecec33265f7eeb6109dc01715f94a0eb6a13149"} Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.764753 4872 scope.go:117] "RemoveContainer" containerID="44cea19e0a7c47a6453c75eb2db092609a4d55ca7c7816f3b2d46300f9bf7181" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.764892 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c4cd99bf6-6gmxf" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.805720 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.827975 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c4cd99bf6-6gmxf"] Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.844441 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5c4cd99bf6-6gmxf"] Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.868915 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d9c67aaa-15db-4359-b978-685a7ed15ec4","Type":"ContainerStarted","Data":"c61462eda146d9cb411719dc7371e033cf76648d5295154f2569e699d0cadbad"} Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.895095 4872 scope.go:117] "RemoveContainer" containerID="269f313a32b515484c32ad8ae94fc4198626212900df162c04dcd127a11599fd" Jan 26 09:24:39 crc kubenswrapper[4872]: I0126 09:24:39.943648 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.067841 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-ovndb-tls-certs\") pod \"ab2bb608-c691-45ba-a23b-b0df661b68b7\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.068346 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-combined-ca-bundle\") pod \"ab2bb608-c691-45ba-a23b-b0df661b68b7\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.068579 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lw5r\" (UniqueName: \"kubernetes.io/projected/ab2bb608-c691-45ba-a23b-b0df661b68b7-kube-api-access-8lw5r\") pod \"ab2bb608-c691-45ba-a23b-b0df661b68b7\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.068605 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-config\") pod \"ab2bb608-c691-45ba-a23b-b0df661b68b7\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.068756 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-httpd-config\") pod \"ab2bb608-c691-45ba-a23b-b0df661b68b7\" (UID: \"ab2bb608-c691-45ba-a23b-b0df661b68b7\") " Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.088137 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "ab2bb608-c691-45ba-a23b-b0df661b68b7" (UID: "ab2bb608-c691-45ba-a23b-b0df661b68b7"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.105503 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab2bb608-c691-45ba-a23b-b0df661b68b7-kube-api-access-8lw5r" (OuterVolumeSpecName: "kube-api-access-8lw5r") pod "ab2bb608-c691-45ba-a23b-b0df661b68b7" (UID: "ab2bb608-c691-45ba-a23b-b0df661b68b7"). InnerVolumeSpecName "kube-api-access-8lw5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.145058 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ab2bb608-c691-45ba-a23b-b0df661b68b7" (UID: "ab2bb608-c691-45ba-a23b-b0df661b68b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.148375 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-config" (OuterVolumeSpecName: "config") pod "ab2bb608-c691-45ba-a23b-b0df661b68b7" (UID: "ab2bb608-c691-45ba-a23b-b0df661b68b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.172672 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lw5r\" (UniqueName: \"kubernetes.io/projected/ab2bb608-c691-45ba-a23b-b0df661b68b7-kube-api-access-8lw5r\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.172954 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.173062 4872 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-httpd-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.173126 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.187916 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "ab2bb608-c691-45ba-a23b-b0df661b68b7" (UID: "ab2bb608-c691-45ba-a23b-b0df661b68b7"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.289165 4872 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab2bb608-c691-45ba-a23b-b0df661b68b7-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.365673 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.855389 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-6785557d58-cspfm" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.925175 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-bc6dfb648-7gm8x"] Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.925664 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ironic-bc6dfb648-7gm8x" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerName="ironic-api-log" containerID="cri-o://a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743" gracePeriod=60 Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.988218 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-88948b444-cjfbr" event={"ID":"ab2bb608-c691-45ba-a23b-b0df661b68b7","Type":"ContainerDied","Data":"46be6e2428d70aa65c521eeb69e2f2cbc70afa7c2612d424729f2d1898fa873a"} Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.988287 4872 scope.go:117] "RemoveContainer" containerID="413e857e7cc2d427c4e6329b01cc361aec394e2fab0a6588eed164991c76e0f0" Jan 26 09:24:40 crc kubenswrapper[4872]: I0126 09:24:40.988498 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-88948b444-cjfbr" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.000576 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"52a92868-e230-4403-9297-51c2b457343b","Type":"ContainerStarted","Data":"80668a7490bf6f689e31b85905c8cc3501d04759a4e8ff23eb17ff171da97456"} Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.006318 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6faf686b-43c3-44f2-9e84-a142ce8192d1","Type":"ContainerStarted","Data":"8c6888e34b5dcd1861cfb7710b9f35366f02b6874339d30832b534a68a606928"} Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.032516 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d9c67aaa-15db-4359-b978-685a7ed15ec4","Type":"ContainerStarted","Data":"d5b9c73f2a178945b99ddd7a6d600ada3b4e32616808cf6f3c36d0ccbd33ab4e"} Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.033636 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.057954 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-88948b444-cjfbr"] Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.069279 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-88948b444-cjfbr"] Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.088124 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=7.08809483 podStartE2EDuration="7.08809483s" podCreationTimestamp="2026-01-26 09:24:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:41.079849668 +0000 UTC m=+1014.388689479" watchObservedRunningTime="2026-01-26 09:24:41.08809483 +0000 UTC m=+1014.396934631" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.117733 4872 scope.go:117] "RemoveContainer" containerID="ef2ebf6d3a564ff492b2caf68378c294a29edd9965cc00af1dc8b28b36f98d7d" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.225341 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c05094d-a3b9-4016-88c1-2eabf12f3920" path="/var/lib/kubelet/pods/5c05094d-a3b9-4016-88c1-2eabf12f3920/volumes" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.226336 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab2bb608-c691-45ba-a23b-b0df661b68b7" path="/var/lib/kubelet/pods/ab2bb608-c691-45ba-a23b-b0df661b68b7/volumes" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.620347 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.636907 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data\") pod \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.637014 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4lz4\" (UniqueName: \"kubernetes.io/projected/2b1a1c03-c8bb-44ab-b607-722109c01b5c-kube-api-access-s4lz4\") pod \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.637088 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b1a1c03-c8bb-44ab-b607-722109c01b5c-logs\") pod \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.637165 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-combined-ca-bundle\") pod \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.637213 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data-custom\") pod \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.637261 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-scripts\") pod \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.637293 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/2b1a1c03-c8bb-44ab-b607-722109c01b5c-etc-podinfo\") pod \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.637393 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data-merged\") pod \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\" (UID: \"2b1a1c03-c8bb-44ab-b607-722109c01b5c\") " Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.637754 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b1a1c03-c8bb-44ab-b607-722109c01b5c-logs" (OuterVolumeSpecName: "logs") pod "2b1a1c03-c8bb-44ab-b607-722109c01b5c" (UID: "2b1a1c03-c8bb-44ab-b607-722109c01b5c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.637876 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b1a1c03-c8bb-44ab-b607-722109c01b5c-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.638396 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "2b1a1c03-c8bb-44ab-b607-722109c01b5c" (UID: "2b1a1c03-c8bb-44ab-b607-722109c01b5c"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.646966 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b1a1c03-c8bb-44ab-b607-722109c01b5c-kube-api-access-s4lz4" (OuterVolumeSpecName: "kube-api-access-s4lz4") pod "2b1a1c03-c8bb-44ab-b607-722109c01b5c" (UID: "2b1a1c03-c8bb-44ab-b607-722109c01b5c"). InnerVolumeSpecName "kube-api-access-s4lz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.651283 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-scripts" (OuterVolumeSpecName: "scripts") pod "2b1a1c03-c8bb-44ab-b607-722109c01b5c" (UID: "2b1a1c03-c8bb-44ab-b607-722109c01b5c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.701099 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2b1a1c03-c8bb-44ab-b607-722109c01b5c" (UID: "2b1a1c03-c8bb-44ab-b607-722109c01b5c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.739625 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2b1a1c03-c8bb-44ab-b607-722109c01b5c-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "2b1a1c03-c8bb-44ab-b607-722109c01b5c" (UID: "2b1a1c03-c8bb-44ab-b607-722109c01b5c"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.742057 4872 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.742090 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.742104 4872 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/2b1a1c03-c8bb-44ab-b607-722109c01b5c-etc-podinfo\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.742142 4872 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data-merged\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.742161 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4lz4\" (UniqueName: \"kubernetes.io/projected/2b1a1c03-c8bb-44ab-b607-722109c01b5c-kube-api-access-s4lz4\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.756464 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data" (OuterVolumeSpecName: "config-data") pod "2b1a1c03-c8bb-44ab-b607-722109c01b5c" (UID: "2b1a1c03-c8bb-44ab-b607-722109c01b5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.785350 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b1a1c03-c8bb-44ab-b607-722109c01b5c" (UID: "2b1a1c03-c8bb-44ab-b607-722109c01b5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.843745 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:41 crc kubenswrapper[4872]: I0126 09:24:41.843823 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b1a1c03-c8bb-44ab-b607-722109c01b5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.053338 4872 generic.go:334] "Generic (PLEG): container finished" podID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerID="a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743" exitCode=143 Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.053452 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-bc6dfb648-7gm8x" Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.053460 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bc6dfb648-7gm8x" event={"ID":"2b1a1c03-c8bb-44ab-b607-722109c01b5c","Type":"ContainerDied","Data":"a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743"} Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.053529 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-bc6dfb648-7gm8x" event={"ID":"2b1a1c03-c8bb-44ab-b607-722109c01b5c","Type":"ContainerDied","Data":"5a36b336e8d85864c8e9db71848fedf896327122bdf23e72611fd1bef6a14a96"} Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.053556 4872 scope.go:117] "RemoveContainer" containerID="726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b" Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.129350 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.164987 4872 scope.go:117] "RemoveContainer" containerID="a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743" Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.176118 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-bc6dfb648-7gm8x"] Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.193288 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-bc6dfb648-7gm8x"] Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.201036 4872 scope.go:117] "RemoveContainer" containerID="0cfb2006217b2ca01dba2c460e11f24eb4ef47861323bd627c96dc867972de97" Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.250086 4872 scope.go:117] "RemoveContainer" containerID="726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b" Jan 26 09:24:42 crc kubenswrapper[4872]: E0126 09:24:42.257490 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b\": container with ID starting with 726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b not found: ID does not exist" containerID="726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b" Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.257541 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b"} err="failed to get container status \"726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b\": rpc error: code = NotFound desc = could not find container \"726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b\": container with ID starting with 726c6cd5ac7253f16a0b1fa86cdf2e9f4938326dded794b585172d0b4cf4c42b not found: ID does not exist" Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.257578 4872 scope.go:117] "RemoveContainer" containerID="a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743" Jan 26 09:24:42 crc kubenswrapper[4872]: E0126 09:24:42.258565 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743\": container with ID starting with a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743 not found: ID does not exist" containerID="a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743" Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.258590 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743"} err="failed to get container status \"a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743\": rpc error: code = NotFound desc = could not find container \"a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743\": container with ID starting with a2c053f40b711751c57efb72a70472f57937e8406eb4d38f62cb7224076ba743 not found: ID does not exist" Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.258605 4872 scope.go:117] "RemoveContainer" containerID="0cfb2006217b2ca01dba2c460e11f24eb4ef47861323bd627c96dc867972de97" Jan 26 09:24:42 crc kubenswrapper[4872]: E0126 09:24:42.259588 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cfb2006217b2ca01dba2c460e11f24eb4ef47861323bd627c96dc867972de97\": container with ID starting with 0cfb2006217b2ca01dba2c460e11f24eb4ef47861323bd627c96dc867972de97 not found: ID does not exist" containerID="0cfb2006217b2ca01dba2c460e11f24eb4ef47861323bd627c96dc867972de97" Jan 26 09:24:42 crc kubenswrapper[4872]: I0126 09:24:42.259639 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cfb2006217b2ca01dba2c460e11f24eb4ef47861323bd627c96dc867972de97"} err="failed to get container status \"0cfb2006217b2ca01dba2c460e11f24eb4ef47861323bd627c96dc867972de97\": rpc error: code = NotFound desc = could not find container \"0cfb2006217b2ca01dba2c460e11f24eb4ef47861323bd627c96dc867972de97\": container with ID starting with 0cfb2006217b2ca01dba2c460e11f24eb4ef47861323bd627c96dc867972de97 not found: ID does not exist" Jan 26 09:24:43 crc kubenswrapper[4872]: I0126 09:24:43.069481 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6faf686b-43c3-44f2-9e84-a142ce8192d1","Type":"ContainerStarted","Data":"fc880a061465bda9031475fc46e469f58b56bf2f94b8b30a0515fe76465f20ee"} Jan 26 09:24:43 crc kubenswrapper[4872]: I0126 09:24:43.073122 4872 generic.go:334] "Generic (PLEG): container finished" podID="f721aeee-7daf-413d-885b-5c146881eb99" containerID="4e2e1a239f503efca62c57a52ec0ca4f72685e7934b746cee1b12681d3e89b09" exitCode=1 Jan 26 09:24:43 crc kubenswrapper[4872]: I0126 09:24:43.074133 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" event={"ID":"f721aeee-7daf-413d-885b-5c146881eb99","Type":"ContainerDied","Data":"4e2e1a239f503efca62c57a52ec0ca4f72685e7934b746cee1b12681d3e89b09"} Jan 26 09:24:43 crc kubenswrapper[4872]: I0126 09:24:43.074201 4872 scope.go:117] "RemoveContainer" containerID="6f4e993edf18dbf2c209d16d44199aa390f4076942602556b0bd58f3cffb4f1e" Jan 26 09:24:43 crc kubenswrapper[4872]: I0126 09:24:43.074504 4872 scope.go:117] "RemoveContainer" containerID="4e2e1a239f503efca62c57a52ec0ca4f72685e7934b746cee1b12681d3e89b09" Jan 26 09:24:43 crc kubenswrapper[4872]: E0126 09:24:43.074731 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-5b47d6fc7b-7wkw2_openstack(f721aeee-7daf-413d-885b-5c146881eb99)\"" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" podUID="f721aeee-7daf-413d-885b-5c146881eb99" Jan 26 09:24:43 crc kubenswrapper[4872]: I0126 09:24:43.153760 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:43 crc kubenswrapper[4872]: I0126 09:24:43.197556 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" path="/var/lib/kubelet/pods/2b1a1c03-c8bb-44ab-b607-722109c01b5c/volumes" Jan 26 09:24:43 crc kubenswrapper[4872]: I0126 09:24:43.316962 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-75b468fcb8-mdk2p" Jan 26 09:24:43 crc kubenswrapper[4872]: I0126 09:24:43.399816 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-78b87b9446-n4fxp"] Jan 26 09:24:43 crc kubenswrapper[4872]: I0126 09:24:43.400204 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-78b87b9446-n4fxp" podUID="421c2c04-702f-403f-a6b0-fa784dd1af3f" containerName="barbican-api-log" containerID="cri-o://34890775e3d16135eed7f885879c0a9df37767fe57fd212923ef600d38b6a25d" gracePeriod=30 Jan 26 09:24:43 crc kubenswrapper[4872]: I0126 09:24:43.400856 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-78b87b9446-n4fxp" podUID="421c2c04-702f-403f-a6b0-fa784dd1af3f" containerName="barbican-api" containerID="cri-o://08cca30da0950126763615468a8e5086c415dffac4f3901edc196c07e58e1f86" gracePeriod=30 Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.146524 4872 generic.go:334] "Generic (PLEG): container finished" podID="421c2c04-702f-403f-a6b0-fa784dd1af3f" containerID="34890775e3d16135eed7f885879c0a9df37767fe57fd212923ef600d38b6a25d" exitCode=143 Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.146703 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78b87b9446-n4fxp" event={"ID":"421c2c04-702f-403f-a6b0-fa784dd1af3f","Type":"ContainerDied","Data":"34890775e3d16135eed7f885879c0a9df37767fe57fd212923ef600d38b6a25d"} Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.576356 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.576931 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="ceilometer-central-agent" containerID="cri-o://155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486" gracePeriod=30 Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.577388 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="proxy-httpd" containerID="cri-o://6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46" gracePeriod=30 Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.577587 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="sg-core" containerID="cri-o://b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90" gracePeriod=30 Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.577631 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="ceilometer-notification-agent" containerID="cri-o://51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54" gracePeriod=30 Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.587516 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.972019 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-68bdb4f9bc-lg4kg"] Jan 26 09:24:44 crc kubenswrapper[4872]: E0126 09:24:44.973197 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerName="ironic-api" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.973219 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerName="ironic-api" Jan 26 09:24:44 crc kubenswrapper[4872]: E0126 09:24:44.973236 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab2bb608-c691-45ba-a23b-b0df661b68b7" containerName="neutron-httpd" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.973244 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab2bb608-c691-45ba-a23b-b0df661b68b7" containerName="neutron-httpd" Jan 26 09:24:44 crc kubenswrapper[4872]: E0126 09:24:44.973263 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerName="ironic-api-log" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.973269 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerName="ironic-api-log" Jan 26 09:24:44 crc kubenswrapper[4872]: E0126 09:24:44.973281 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerName="ironic-api" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.973287 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerName="ironic-api" Jan 26 09:24:44 crc kubenswrapper[4872]: E0126 09:24:44.973303 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerName="init" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.973310 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerName="init" Jan 26 09:24:44 crc kubenswrapper[4872]: E0126 09:24:44.973324 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab2bb608-c691-45ba-a23b-b0df661b68b7" containerName="neutron-api" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.973330 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab2bb608-c691-45ba-a23b-b0df661b68b7" containerName="neutron-api" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.975392 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerName="ironic-api" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.975427 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab2bb608-c691-45ba-a23b-b0df661b68b7" containerName="neutron-httpd" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.975441 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerName="ironic-api-log" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.975449 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab2bb608-c691-45ba-a23b-b0df661b68b7" containerName="neutron-api" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.975771 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b1a1c03-c8bb-44ab-b607-722109c01b5c" containerName="ironic-api" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.976458 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.991352 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.991497 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Jan 26 09:24:44 crc kubenswrapper[4872]: I0126 09:24:44.991695 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.004157 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-68bdb4f9bc-lg4kg"] Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.128102 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15c32f5-7a02-497d-a689-f3c82c63ad8b-config-data\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.128168 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c15c32f5-7a02-497d-a689-f3c82c63ad8b-log-httpd\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.128192 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6gv2\" (UniqueName: \"kubernetes.io/projected/c15c32f5-7a02-497d-a689-f3c82c63ad8b-kube-api-access-f6gv2\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.128224 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c15c32f5-7a02-497d-a689-f3c82c63ad8b-run-httpd\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.128245 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c15c32f5-7a02-497d-a689-f3c82c63ad8b-etc-swift\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.128276 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15c32f5-7a02-497d-a689-f3c82c63ad8b-combined-ca-bundle\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.128292 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c15c32f5-7a02-497d-a689-f3c82c63ad8b-public-tls-certs\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.128321 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c15c32f5-7a02-497d-a689-f3c82c63ad8b-internal-tls-certs\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.211454 4872 generic.go:334] "Generic (PLEG): container finished" podID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerID="6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46" exitCode=0 Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.211492 4872 generic.go:334] "Generic (PLEG): container finished" podID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerID="b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90" exitCode=2 Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.216581 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6d0ffab-33e4-4cce-8517-9d081b31f695","Type":"ContainerDied","Data":"6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46"} Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.216620 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6d0ffab-33e4-4cce-8517-9d081b31f695","Type":"ContainerDied","Data":"b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90"} Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.222642 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-dv876" event={"ID":"909a12a7-4b22-43b8-8d96-1fff1ef2738c","Type":"ContainerStarted","Data":"94993efef6b44ee41cc9a997286d195f3455db8520d20a31d2cc848d9519f899"} Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.229694 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15c32f5-7a02-497d-a689-f3c82c63ad8b-config-data\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.229769 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c15c32f5-7a02-497d-a689-f3c82c63ad8b-log-httpd\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.229813 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6gv2\" (UniqueName: \"kubernetes.io/projected/c15c32f5-7a02-497d-a689-f3c82c63ad8b-kube-api-access-f6gv2\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.229842 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c15c32f5-7a02-497d-a689-f3c82c63ad8b-run-httpd\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.229866 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c15c32f5-7a02-497d-a689-f3c82c63ad8b-etc-swift\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.229903 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15c32f5-7a02-497d-a689-f3c82c63ad8b-combined-ca-bundle\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.229920 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c15c32f5-7a02-497d-a689-f3c82c63ad8b-public-tls-certs\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.229946 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c15c32f5-7a02-497d-a689-f3c82c63ad8b-internal-tls-certs\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.230209 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c15c32f5-7a02-497d-a689-f3c82c63ad8b-log-httpd\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.231295 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c15c32f5-7a02-497d-a689-f3c82c63ad8b-run-httpd\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.244276 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c15c32f5-7a02-497d-a689-f3c82c63ad8b-config-data\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.245625 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c15c32f5-7a02-497d-a689-f3c82c63ad8b-internal-tls-certs\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.247599 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-inspector-db-sync-dv876" podStartSLOduration=3.763159063 podStartE2EDuration="9.247581986s" podCreationTimestamp="2026-01-26 09:24:36 +0000 UTC" firstStartedPulling="2026-01-26 09:24:39.152992296 +0000 UTC m=+1012.461832097" lastFinishedPulling="2026-01-26 09:24:44.637415229 +0000 UTC m=+1017.946255020" observedRunningTime="2026-01-26 09:24:45.237974879 +0000 UTC m=+1018.546814680" watchObservedRunningTime="2026-01-26 09:24:45.247581986 +0000 UTC m=+1018.556421787" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.254659 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c15c32f5-7a02-497d-a689-f3c82c63ad8b-etc-swift\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.255149 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6gv2\" (UniqueName: \"kubernetes.io/projected/c15c32f5-7a02-497d-a689-f3c82c63ad8b-kube-api-access-f6gv2\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.256217 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c15c32f5-7a02-497d-a689-f3c82c63ad8b-public-tls-certs\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.259886 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c15c32f5-7a02-497d-a689-f3c82c63ad8b-combined-ca-bundle\") pod \"swift-proxy-68bdb4f9bc-lg4kg\" (UID: \"c15c32f5-7a02-497d-a689-f3c82c63ad8b\") " pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.326023 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:45 crc kubenswrapper[4872]: I0126 09:24:45.980338 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-68bdb4f9bc-lg4kg"] Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.134281 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.185467 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6d0ffab-33e4-4cce-8517-9d081b31f695-log-httpd\") pod \"f6d0ffab-33e4-4cce-8517-9d081b31f695\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.186004 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-config-data\") pod \"f6d0ffab-33e4-4cce-8517-9d081b31f695\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.186103 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlj7l\" (UniqueName: \"kubernetes.io/projected/f6d0ffab-33e4-4cce-8517-9d081b31f695-kube-api-access-mlj7l\") pod \"f6d0ffab-33e4-4cce-8517-9d081b31f695\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.186295 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-scripts\") pod \"f6d0ffab-33e4-4cce-8517-9d081b31f695\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.186322 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-sg-core-conf-yaml\") pod \"f6d0ffab-33e4-4cce-8517-9d081b31f695\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.186346 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6d0ffab-33e4-4cce-8517-9d081b31f695-run-httpd\") pod \"f6d0ffab-33e4-4cce-8517-9d081b31f695\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.186432 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-combined-ca-bundle\") pod \"f6d0ffab-33e4-4cce-8517-9d081b31f695\" (UID: \"f6d0ffab-33e4-4cce-8517-9d081b31f695\") " Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.188391 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6d0ffab-33e4-4cce-8517-9d081b31f695-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f6d0ffab-33e4-4cce-8517-9d081b31f695" (UID: "f6d0ffab-33e4-4cce-8517-9d081b31f695"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.190372 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6d0ffab-33e4-4cce-8517-9d081b31f695-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f6d0ffab-33e4-4cce-8517-9d081b31f695" (UID: "f6d0ffab-33e4-4cce-8517-9d081b31f695"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.197859 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6d0ffab-33e4-4cce-8517-9d081b31f695-kube-api-access-mlj7l" (OuterVolumeSpecName: "kube-api-access-mlj7l") pod "f6d0ffab-33e4-4cce-8517-9d081b31f695" (UID: "f6d0ffab-33e4-4cce-8517-9d081b31f695"). InnerVolumeSpecName "kube-api-access-mlj7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.199848 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-scripts" (OuterVolumeSpecName: "scripts") pod "f6d0ffab-33e4-4cce-8517-9d081b31f695" (UID: "f6d0ffab-33e4-4cce-8517-9d081b31f695"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.250594 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" event={"ID":"c15c32f5-7a02-497d-a689-f3c82c63ad8b","Type":"ContainerStarted","Data":"04afd5e3da57f102ea292924bd3d561fef312888a8d0abedb34a20d5f55c5257"} Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.264658 4872 generic.go:334] "Generic (PLEG): container finished" podID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerID="51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54" exitCode=0 Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.264694 4872 generic.go:334] "Generic (PLEG): container finished" podID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerID="155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486" exitCode=0 Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.264852 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.265221 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6d0ffab-33e4-4cce-8517-9d081b31f695","Type":"ContainerDied","Data":"51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54"} Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.265271 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6d0ffab-33e4-4cce-8517-9d081b31f695","Type":"ContainerDied","Data":"155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486"} Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.265284 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6d0ffab-33e4-4cce-8517-9d081b31f695","Type":"ContainerDied","Data":"ae7ab075617029cfe3463f3e6cb8e14eeddae87939c4998f8cfb2a396f562e78"} Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.265300 4872 scope.go:117] "RemoveContainer" containerID="6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.266477 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f6d0ffab-33e4-4cce-8517-9d081b31f695" (UID: "f6d0ffab-33e4-4cce-8517-9d081b31f695"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.274830 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6faf686b-43c3-44f2-9e84-a142ce8192d1","Type":"ContainerStarted","Data":"85ff61bf9669bdf75073c8e1d7f78e62180f9cf31fe84a09d1c5f71b6e6c0e8f"} Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.289413 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6d0ffab-33e4-4cce-8517-9d081b31f695-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.289449 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlj7l\" (UniqueName: \"kubernetes.io/projected/f6d0ffab-33e4-4cce-8517-9d081b31f695-kube-api-access-mlj7l\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.289461 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.289471 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.289480 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6d0ffab-33e4-4cce-8517-9d081b31f695-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.300910 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=7.300887059 podStartE2EDuration="7.300887059s" podCreationTimestamp="2026-01-26 09:24:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:46.293991612 +0000 UTC m=+1019.602831403" watchObservedRunningTime="2026-01-26 09:24:46.300887059 +0000 UTC m=+1019.609726860" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.309483 4872 scope.go:117] "RemoveContainer" containerID="b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.329872 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6d0ffab-33e4-4cce-8517-9d081b31f695" (UID: "f6d0ffab-33e4-4cce-8517-9d081b31f695"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.337031 4872 scope.go:117] "RemoveContainer" containerID="51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.385685 4872 scope.go:117] "RemoveContainer" containerID="155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.389307 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-config-data" (OuterVolumeSpecName: "config-data") pod "f6d0ffab-33e4-4cce-8517-9d081b31f695" (UID: "f6d0ffab-33e4-4cce-8517-9d081b31f695"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.390806 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.390914 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d0ffab-33e4-4cce-8517-9d081b31f695-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.439335 4872 scope.go:117] "RemoveContainer" containerID="6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46" Jan 26 09:24:46 crc kubenswrapper[4872]: E0126 09:24:46.439945 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46\": container with ID starting with 6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46 not found: ID does not exist" containerID="6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.439980 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46"} err="failed to get container status \"6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46\": rpc error: code = NotFound desc = could not find container \"6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46\": container with ID starting with 6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46 not found: ID does not exist" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.440015 4872 scope.go:117] "RemoveContainer" containerID="b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90" Jan 26 09:24:46 crc kubenswrapper[4872]: E0126 09:24:46.440337 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90\": container with ID starting with b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90 not found: ID does not exist" containerID="b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.440381 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90"} err="failed to get container status \"b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90\": rpc error: code = NotFound desc = could not find container \"b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90\": container with ID starting with b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90 not found: ID does not exist" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.440415 4872 scope.go:117] "RemoveContainer" containerID="51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54" Jan 26 09:24:46 crc kubenswrapper[4872]: E0126 09:24:46.443884 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54\": container with ID starting with 51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54 not found: ID does not exist" containerID="51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.443913 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54"} err="failed to get container status \"51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54\": rpc error: code = NotFound desc = could not find container \"51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54\": container with ID starting with 51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54 not found: ID does not exist" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.443927 4872 scope.go:117] "RemoveContainer" containerID="155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486" Jan 26 09:24:46 crc kubenswrapper[4872]: E0126 09:24:46.444254 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486\": container with ID starting with 155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486 not found: ID does not exist" containerID="155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.444274 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486"} err="failed to get container status \"155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486\": rpc error: code = NotFound desc = could not find container \"155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486\": container with ID starting with 155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486 not found: ID does not exist" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.444286 4872 scope.go:117] "RemoveContainer" containerID="6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.445044 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46"} err="failed to get container status \"6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46\": rpc error: code = NotFound desc = could not find container \"6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46\": container with ID starting with 6d0531df62f15de97e3c0d4232bcac1bade8333b51983445ea3bab4f83a56c46 not found: ID does not exist" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.445066 4872 scope.go:117] "RemoveContainer" containerID="b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.445379 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90"} err="failed to get container status \"b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90\": rpc error: code = NotFound desc = could not find container \"b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90\": container with ID starting with b1d2b09ebbb648b3b59982a29bf1ce7a9b026e6e97421f3e89af19a0714c6f90 not found: ID does not exist" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.445404 4872 scope.go:117] "RemoveContainer" containerID="51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.447534 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54"} err="failed to get container status \"51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54\": rpc error: code = NotFound desc = could not find container \"51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54\": container with ID starting with 51ddf669d71e2a32578ae440cd3e5e5bce70f21d09cb0fb353abcb3e57565d54 not found: ID does not exist" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.447560 4872 scope.go:117] "RemoveContainer" containerID="155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.448285 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486"} err="failed to get container status \"155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486\": rpc error: code = NotFound desc = could not find container \"155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486\": container with ID starting with 155f230a03b28496e4c221dbfcdbe44fce676a31e151502a777fdf931712a486 not found: ID does not exist" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.611629 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.622577 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.633499 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:24:46 crc kubenswrapper[4872]: E0126 09:24:46.634041 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="ceilometer-notification-agent" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.634063 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="ceilometer-notification-agent" Jan 26 09:24:46 crc kubenswrapper[4872]: E0126 09:24:46.634092 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="proxy-httpd" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.634099 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="proxy-httpd" Jan 26 09:24:46 crc kubenswrapper[4872]: E0126 09:24:46.634117 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="ceilometer-central-agent" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.634123 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="ceilometer-central-agent" Jan 26 09:24:46 crc kubenswrapper[4872]: E0126 09:24:46.634132 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="sg-core" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.634138 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="sg-core" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.634393 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="proxy-httpd" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.634412 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="sg-core" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.634426 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="ceilometer-central-agent" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.634454 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" containerName="ceilometer-notification-agent" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.636517 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.640859 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.641147 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.670370 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78b87b9446-n4fxp" podUID="421c2c04-702f-403f-a6b0-fa784dd1af3f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:32880->10.217.0.160:9311: read: connection reset by peer" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.670690 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78b87b9446-n4fxp" podUID="421c2c04-702f-403f-a6b0-fa784dd1af3f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:32866->10.217.0.160:9311: read: connection reset by peer" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.811097 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.811168 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94a8957b-b742-4dfd-a463-e265cf4db32e-log-httpd\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.811242 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l5p2\" (UniqueName: \"kubernetes.io/projected/94a8957b-b742-4dfd-a463-e265cf4db32e-kube-api-access-6l5p2\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.811273 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94a8957b-b742-4dfd-a463-e265cf4db32e-run-httpd\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.811309 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-scripts\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.811326 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.811346 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-config-data\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.814916 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.913197 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.913667 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94a8957b-b742-4dfd-a463-e265cf4db32e-log-httpd\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.913726 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l5p2\" (UniqueName: \"kubernetes.io/projected/94a8957b-b742-4dfd-a463-e265cf4db32e-kube-api-access-6l5p2\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.913760 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94a8957b-b742-4dfd-a463-e265cf4db32e-run-httpd\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.913776 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-scripts\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.913791 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.913822 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-config-data\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.914419 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94a8957b-b742-4dfd-a463-e265cf4db32e-log-httpd\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.914437 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94a8957b-b742-4dfd-a463-e265cf4db32e-run-httpd\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.920365 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.922573 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-config-data\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.924208 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.925540 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-scripts\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:46 crc kubenswrapper[4872]: I0126 09:24:46.934043 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l5p2\" (UniqueName: \"kubernetes.io/projected/94a8957b-b742-4dfd-a463-e265cf4db32e-kube-api-access-6l5p2\") pod \"ceilometer-0\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " pod="openstack/ceilometer-0" Jan 26 09:24:47 crc kubenswrapper[4872]: I0126 09:24:47.073845 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:47 crc kubenswrapper[4872]: I0126 09:24:47.073928 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:24:47 crc kubenswrapper[4872]: I0126 09:24:47.074784 4872 scope.go:117] "RemoveContainer" containerID="4e2e1a239f503efca62c57a52ec0ca4f72685e7934b746cee1b12681d3e89b09" Jan 26 09:24:47 crc kubenswrapper[4872]: E0126 09:24:47.075077 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-5b47d6fc7b-7wkw2_openstack(f721aeee-7daf-413d-885b-5c146881eb99)\"" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" podUID="f721aeee-7daf-413d-885b-5c146881eb99" Jan 26 09:24:47 crc kubenswrapper[4872]: I0126 09:24:47.194958 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:24:47 crc kubenswrapper[4872]: I0126 09:24:47.200572 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6d0ffab-33e4-4cce-8517-9d081b31f695" path="/var/lib/kubelet/pods/f6d0ffab-33e4-4cce-8517-9d081b31f695/volumes" Jan 26 09:24:47 crc kubenswrapper[4872]: I0126 09:24:47.304596 4872 generic.go:334] "Generic (PLEG): container finished" podID="421c2c04-702f-403f-a6b0-fa784dd1af3f" containerID="08cca30da0950126763615468a8e5086c415dffac4f3901edc196c07e58e1f86" exitCode=0 Jan 26 09:24:47 crc kubenswrapper[4872]: I0126 09:24:47.304813 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78b87b9446-n4fxp" event={"ID":"421c2c04-702f-403f-a6b0-fa784dd1af3f","Type":"ContainerDied","Data":"08cca30da0950126763615468a8e5086c415dffac4f3901edc196c07e58e1f86"} Jan 26 09:24:47 crc kubenswrapper[4872]: I0126 09:24:47.324366 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" event={"ID":"c15c32f5-7a02-497d-a689-f3c82c63ad8b","Type":"ContainerStarted","Data":"74080d02e5f58040449bec91817e11199077c5604c9a40512d876d46761bc4a3"} Jan 26 09:24:47 crc kubenswrapper[4872]: I0126 09:24:47.597599 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Jan 26 09:24:47 crc kubenswrapper[4872]: I0126 09:24:47.720231 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:24:47 crc kubenswrapper[4872]: W0126 09:24:47.734502 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94a8957b_b742_4dfd_a463_e265cf4db32e.slice/crio-8e256f948be1dfdc015f44d318ec72c7e2c3b157c8c0e5b7bbad072f0c919a9d WatchSource:0}: Error finding container 8e256f948be1dfdc015f44d318ec72c7e2c3b157c8c0e5b7bbad072f0c919a9d: Status 404 returned error can't find the container with id 8e256f948be1dfdc015f44d318ec72c7e2c3b157c8c0e5b7bbad072f0c919a9d Jan 26 09:24:47 crc kubenswrapper[4872]: I0126 09:24:47.878500 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.058971 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-combined-ca-bundle\") pod \"421c2c04-702f-403f-a6b0-fa784dd1af3f\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.059083 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tflrm\" (UniqueName: \"kubernetes.io/projected/421c2c04-702f-403f-a6b0-fa784dd1af3f-kube-api-access-tflrm\") pod \"421c2c04-702f-403f-a6b0-fa784dd1af3f\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.059171 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/421c2c04-702f-403f-a6b0-fa784dd1af3f-logs\") pod \"421c2c04-702f-403f-a6b0-fa784dd1af3f\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.059211 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data\") pod \"421c2c04-702f-403f-a6b0-fa784dd1af3f\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.059346 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data-custom\") pod \"421c2c04-702f-403f-a6b0-fa784dd1af3f\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.060831 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/421c2c04-702f-403f-a6b0-fa784dd1af3f-logs" (OuterVolumeSpecName: "logs") pod "421c2c04-702f-403f-a6b0-fa784dd1af3f" (UID: "421c2c04-702f-403f-a6b0-fa784dd1af3f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.070133 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/421c2c04-702f-403f-a6b0-fa784dd1af3f-kube-api-access-tflrm" (OuterVolumeSpecName: "kube-api-access-tflrm") pod "421c2c04-702f-403f-a6b0-fa784dd1af3f" (UID: "421c2c04-702f-403f-a6b0-fa784dd1af3f"). InnerVolumeSpecName "kube-api-access-tflrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.076965 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "421c2c04-702f-403f-a6b0-fa784dd1af3f" (UID: "421c2c04-702f-403f-a6b0-fa784dd1af3f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.099207 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "421c2c04-702f-403f-a6b0-fa784dd1af3f" (UID: "421c2c04-702f-403f-a6b0-fa784dd1af3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.167018 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data" (OuterVolumeSpecName: "config-data") pod "421c2c04-702f-403f-a6b0-fa784dd1af3f" (UID: "421c2c04-702f-403f-a6b0-fa784dd1af3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.169586 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data\") pod \"421c2c04-702f-403f-a6b0-fa784dd1af3f\" (UID: \"421c2c04-702f-403f-a6b0-fa784dd1af3f\") " Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.170628 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tflrm\" (UniqueName: \"kubernetes.io/projected/421c2c04-702f-403f-a6b0-fa784dd1af3f-kube-api-access-tflrm\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.170649 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/421c2c04-702f-403f-a6b0-fa784dd1af3f-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.170661 4872 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data-custom\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.170672 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:48 crc kubenswrapper[4872]: W0126 09:24:48.170792 4872 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/421c2c04-702f-403f-a6b0-fa784dd1af3f/volumes/kubernetes.io~secret/config-data Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.170838 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data" (OuterVolumeSpecName: "config-data") pod "421c2c04-702f-403f-a6b0-fa784dd1af3f" (UID: "421c2c04-702f-403f-a6b0-fa784dd1af3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.273702 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/421c2c04-702f-403f-a6b0-fa784dd1af3f-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.336850 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" event={"ID":"c15c32f5-7a02-497d-a689-f3c82c63ad8b","Type":"ContainerStarted","Data":"f63368ffeb3e5fd02c5854df5d2f98859a20f5e8ac0f8a25c070b7bcd3ab403a"} Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.336964 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.336995 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.340376 4872 generic.go:334] "Generic (PLEG): container finished" podID="909a12a7-4b22-43b8-8d96-1fff1ef2738c" containerID="94993efef6b44ee41cc9a997286d195f3455db8520d20a31d2cc848d9519f899" exitCode=0 Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.340505 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-dv876" event={"ID":"909a12a7-4b22-43b8-8d96-1fff1ef2738c","Type":"ContainerDied","Data":"94993efef6b44ee41cc9a997286d195f3455db8520d20a31d2cc848d9519f899"} Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.345610 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94a8957b-b742-4dfd-a463-e265cf4db32e","Type":"ContainerStarted","Data":"8e256f948be1dfdc015f44d318ec72c7e2c3b157c8c0e5b7bbad072f0c919a9d"} Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.347577 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78b87b9446-n4fxp" event={"ID":"421c2c04-702f-403f-a6b0-fa784dd1af3f","Type":"ContainerDied","Data":"fdb297fa0288c7141bb3c4724a52271c2951338a799be5ea60540e97b40f5269"} Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.347616 4872 scope.go:117] "RemoveContainer" containerID="08cca30da0950126763615468a8e5086c415dffac4f3901edc196c07e58e1f86" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.347765 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78b87b9446-n4fxp" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.370339 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" podStartSLOduration=4.370315312 podStartE2EDuration="4.370315312s" podCreationTimestamp="2026-01-26 09:24:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:24:48.354367612 +0000 UTC m=+1021.663207413" watchObservedRunningTime="2026-01-26 09:24:48.370315312 +0000 UTC m=+1021.679155113" Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.412836 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-78b87b9446-n4fxp"] Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.422181 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-78b87b9446-n4fxp"] Jan 26 09:24:48 crc kubenswrapper[4872]: I0126 09:24:48.428598 4872 scope.go:117] "RemoveContainer" containerID="34890775e3d16135eed7f885879c0a9df37767fe57fd212923ef600d38b6a25d" Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.198837 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="421c2c04-702f-403f-a6b0-fa784dd1af3f" path="/var/lib/kubelet/pods/421c2c04-702f-403f-a6b0-fa784dd1af3f/volumes" Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.375283 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94a8957b-b742-4dfd-a463-e265cf4db32e","Type":"ContainerStarted","Data":"2272551f1062c8f5581604bfa484ed493b479d73101aa07319ce08c2141898ee"} Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.666400 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.820786 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.929114 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/909a12a7-4b22-43b8-8d96-1fff1ef2738c-var-lib-ironic\") pod \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.929221 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-config\") pod \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.929293 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-scripts\") pod \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.929325 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ntc4\" (UniqueName: \"kubernetes.io/projected/909a12a7-4b22-43b8-8d96-1fff1ef2738c-kube-api-access-8ntc4\") pod \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.929430 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-combined-ca-bundle\") pod \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.929515 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/909a12a7-4b22-43b8-8d96-1fff1ef2738c-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.929582 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/909a12a7-4b22-43b8-8d96-1fff1ef2738c-etc-podinfo\") pod \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\" (UID: \"909a12a7-4b22-43b8-8d96-1fff1ef2738c\") " Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.929829 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/909a12a7-4b22-43b8-8d96-1fff1ef2738c-var-lib-ironic" (OuterVolumeSpecName: "var-lib-ironic") pod "909a12a7-4b22-43b8-8d96-1fff1ef2738c" (UID: "909a12a7-4b22-43b8-8d96-1fff1ef2738c"). InnerVolumeSpecName "var-lib-ironic". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.930305 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/909a12a7-4b22-43b8-8d96-1fff1ef2738c-var-lib-ironic-inspector-dhcp-hostsdir" (OuterVolumeSpecName: "var-lib-ironic-inspector-dhcp-hostsdir") pod "909a12a7-4b22-43b8-8d96-1fff1ef2738c" (UID: "909a12a7-4b22-43b8-8d96-1fff1ef2738c"). InnerVolumeSpecName "var-lib-ironic-inspector-dhcp-hostsdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.936977 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-scripts" (OuterVolumeSpecName: "scripts") pod "909a12a7-4b22-43b8-8d96-1fff1ef2738c" (UID: "909a12a7-4b22-43b8-8d96-1fff1ef2738c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.940516 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.941123 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/909a12a7-4b22-43b8-8d96-1fff1ef2738c-kube-api-access-8ntc4" (OuterVolumeSpecName: "kube-api-access-8ntc4") pod "909a12a7-4b22-43b8-8d96-1fff1ef2738c" (UID: "909a12a7-4b22-43b8-8d96-1fff1ef2738c"). InnerVolumeSpecName "kube-api-access-8ntc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:24:49 crc kubenswrapper[4872]: I0126 09:24:49.953073 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/909a12a7-4b22-43b8-8d96-1fff1ef2738c-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "909a12a7-4b22-43b8-8d96-1fff1ef2738c" (UID: "909a12a7-4b22-43b8-8d96-1fff1ef2738c"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.010054 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-config" (OuterVolumeSpecName: "config") pod "909a12a7-4b22-43b8-8d96-1fff1ef2738c" (UID: "909a12a7-4b22-43b8-8d96-1fff1ef2738c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.034251 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.034289 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.034299 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ntc4\" (UniqueName: \"kubernetes.io/projected/909a12a7-4b22-43b8-8d96-1fff1ef2738c-kube-api-access-8ntc4\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.034311 4872 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/909a12a7-4b22-43b8-8d96-1fff1ef2738c-var-lib-ironic-inspector-dhcp-hostsdir\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.034321 4872 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/909a12a7-4b22-43b8-8d96-1fff1ef2738c-etc-podinfo\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.034330 4872 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/909a12a7-4b22-43b8-8d96-1fff1ef2738c-var-lib-ironic\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.075899 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "909a12a7-4b22-43b8-8d96-1fff1ef2738c" (UID: "909a12a7-4b22-43b8-8d96-1fff1ef2738c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.136045 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/909a12a7-4b22-43b8-8d96-1fff1ef2738c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.393694 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-dv876" event={"ID":"909a12a7-4b22-43b8-8d96-1fff1ef2738c","Type":"ContainerDied","Data":"adc796bea19110194b6372ef1b76d1f821128af90d2b05fc61b80a2bc3866580"} Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.393812 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adc796bea19110194b6372ef1b76d1f821128af90d2b05fc61b80a2bc3866580" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.393908 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-dv876" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.413871 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94a8957b-b742-4dfd-a463-e265cf4db32e","Type":"ContainerStarted","Data":"f850eae1ba4ede05658e235af04af5e9ddc091ac135bdddf5732952720385fd6"} Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.554683 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-0"] Jan 26 09:24:50 crc kubenswrapper[4872]: E0126 09:24:50.555336 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="909a12a7-4b22-43b8-8d96-1fff1ef2738c" containerName="ironic-inspector-db-sync" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.555414 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="909a12a7-4b22-43b8-8d96-1fff1ef2738c" containerName="ironic-inspector-db-sync" Jan 26 09:24:50 crc kubenswrapper[4872]: E0126 09:24:50.555490 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="421c2c04-702f-403f-a6b0-fa784dd1af3f" containerName="barbican-api-log" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.555545 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="421c2c04-702f-403f-a6b0-fa784dd1af3f" containerName="barbican-api-log" Jan 26 09:24:50 crc kubenswrapper[4872]: E0126 09:24:50.555603 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="421c2c04-702f-403f-a6b0-fa784dd1af3f" containerName="barbican-api" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.555656 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="421c2c04-702f-403f-a6b0-fa784dd1af3f" containerName="barbican-api" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.555900 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="421c2c04-702f-403f-a6b0-fa784dd1af3f" containerName="barbican-api-log" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.555980 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="421c2c04-702f-403f-a6b0-fa784dd1af3f" containerName="barbican-api" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.556043 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="909a12a7-4b22-43b8-8d96-1fff1ef2738c" containerName="ironic-inspector-db-sync" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.559573 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.565280 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.566015 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.574079 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.765501 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nff8\" (UniqueName: \"kubernetes.io/projected/f8a676a6-3108-4171-82c0-9f7c08e521f8-kube-api-access-2nff8\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.765557 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.765617 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/f8a676a6-3108-4171-82c0-9f7c08e521f8-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.765649 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-config\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.765683 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f8a676a6-3108-4171-82c0-9f7c08e521f8-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.765700 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-scripts\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.765718 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/f8a676a6-3108-4171-82c0-9f7c08e521f8-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.867247 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nff8\" (UniqueName: \"kubernetes.io/projected/f8a676a6-3108-4171-82c0-9f7c08e521f8-kube-api-access-2nff8\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.867622 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.867678 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/f8a676a6-3108-4171-82c0-9f7c08e521f8-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.867711 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-config\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.867742 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f8a676a6-3108-4171-82c0-9f7c08e521f8-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.867758 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-scripts\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.867775 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/f8a676a6-3108-4171-82c0-9f7c08e521f8-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.868298 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/f8a676a6-3108-4171-82c0-9f7c08e521f8-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.871082 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/f8a676a6-3108-4171-82c0-9f7c08e521f8-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.872638 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-scripts\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.876593 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f8a676a6-3108-4171-82c0-9f7c08e521f8-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.878553 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-config\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.892637 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nff8\" (UniqueName: \"kubernetes.io/projected/f8a676a6-3108-4171-82c0-9f7c08e521f8-kube-api-access-2nff8\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.896569 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " pod="openstack/ironic-inspector-0" Jan 26 09:24:50 crc kubenswrapper[4872]: I0126 09:24:50.959547 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Jan 26 09:24:52 crc kubenswrapper[4872]: I0126 09:24:52.527872 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.246597 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-0"] Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.495219 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-6dpv7"] Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.496447 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6dpv7" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.513306 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-6dpv7"] Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.599007 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-cl2gt"] Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.600399 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cl2gt" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.606763 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cl2gt"] Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.626226 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sl5x\" (UniqueName: \"kubernetes.io/projected/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff-kube-api-access-2sl5x\") pod \"nova-api-db-create-6dpv7\" (UID: \"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff\") " pod="openstack/nova-api-db-create-6dpv7" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.626359 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff-operator-scripts\") pod \"nova-api-db-create-6dpv7\" (UID: \"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff\") " pod="openstack/nova-api-db-create-6dpv7" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.708007 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-f91e-account-create-update-twns5"] Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.709395 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f91e-account-create-update-twns5" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.713153 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.728070 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sl5x\" (UniqueName: \"kubernetes.io/projected/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff-kube-api-access-2sl5x\") pod \"nova-api-db-create-6dpv7\" (UID: \"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff\") " pod="openstack/nova-api-db-create-6dpv7" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.728164 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f99a167-975f-4373-8c9d-23a2ab6510fa-operator-scripts\") pod \"nova-cell0-db-create-cl2gt\" (UID: \"4f99a167-975f-4373-8c9d-23a2ab6510fa\") " pod="openstack/nova-cell0-db-create-cl2gt" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.728230 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgb9k\" (UniqueName: \"kubernetes.io/projected/4f99a167-975f-4373-8c9d-23a2ab6510fa-kube-api-access-fgb9k\") pod \"nova-cell0-db-create-cl2gt\" (UID: \"4f99a167-975f-4373-8c9d-23a2ab6510fa\") " pod="openstack/nova-cell0-db-create-cl2gt" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.728266 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff-operator-scripts\") pod \"nova-api-db-create-6dpv7\" (UID: \"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff\") " pod="openstack/nova-api-db-create-6dpv7" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.729633 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff-operator-scripts\") pod \"nova-api-db-create-6dpv7\" (UID: \"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff\") " pod="openstack/nova-api-db-create-6dpv7" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.735958 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f91e-account-create-update-twns5"] Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.748288 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sl5x\" (UniqueName: \"kubernetes.io/projected/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff-kube-api-access-2sl5x\") pod \"nova-api-db-create-6dpv7\" (UID: \"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff\") " pod="openstack/nova-api-db-create-6dpv7" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.813246 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-bkbz5"] Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.818009 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6dpv7" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.818261 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bkbz5" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.822570 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bkbz5"] Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.834549 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f99a167-975f-4373-8c9d-23a2ab6510fa-operator-scripts\") pod \"nova-cell0-db-create-cl2gt\" (UID: \"4f99a167-975f-4373-8c9d-23a2ab6510fa\") " pod="openstack/nova-cell0-db-create-cl2gt" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.834630 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jphwx\" (UniqueName: \"kubernetes.io/projected/06fe0f9a-dbc9-4078-92ce-07fe06eb4110-kube-api-access-jphwx\") pod \"nova-api-f91e-account-create-update-twns5\" (UID: \"06fe0f9a-dbc9-4078-92ce-07fe06eb4110\") " pod="openstack/nova-api-f91e-account-create-update-twns5" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.834734 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06fe0f9a-dbc9-4078-92ce-07fe06eb4110-operator-scripts\") pod \"nova-api-f91e-account-create-update-twns5\" (UID: \"06fe0f9a-dbc9-4078-92ce-07fe06eb4110\") " pod="openstack/nova-api-f91e-account-create-update-twns5" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.835018 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgb9k\" (UniqueName: \"kubernetes.io/projected/4f99a167-975f-4373-8c9d-23a2ab6510fa-kube-api-access-fgb9k\") pod \"nova-cell0-db-create-cl2gt\" (UID: \"4f99a167-975f-4373-8c9d-23a2ab6510fa\") " pod="openstack/nova-cell0-db-create-cl2gt" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.836378 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f99a167-975f-4373-8c9d-23a2ab6510fa-operator-scripts\") pod \"nova-cell0-db-create-cl2gt\" (UID: \"4f99a167-975f-4373-8c9d-23a2ab6510fa\") " pod="openstack/nova-cell0-db-create-cl2gt" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.857608 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgb9k\" (UniqueName: \"kubernetes.io/projected/4f99a167-975f-4373-8c9d-23a2ab6510fa-kube-api-access-fgb9k\") pod \"nova-cell0-db-create-cl2gt\" (UID: \"4f99a167-975f-4373-8c9d-23a2ab6510fa\") " pod="openstack/nova-cell0-db-create-cl2gt" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.907840 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-5935-account-create-update-ftjg4"] Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.909094 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5935-account-create-update-ftjg4" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.922345 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cl2gt" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.922530 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.941132 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jphwx\" (UniqueName: \"kubernetes.io/projected/06fe0f9a-dbc9-4078-92ce-07fe06eb4110-kube-api-access-jphwx\") pod \"nova-api-f91e-account-create-update-twns5\" (UID: \"06fe0f9a-dbc9-4078-92ce-07fe06eb4110\") " pod="openstack/nova-api-f91e-account-create-update-twns5" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.941206 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06fe0f9a-dbc9-4078-92ce-07fe06eb4110-operator-scripts\") pod \"nova-api-f91e-account-create-update-twns5\" (UID: \"06fe0f9a-dbc9-4078-92ce-07fe06eb4110\") " pod="openstack/nova-api-f91e-account-create-update-twns5" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.941291 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24282d99-b526-4bd3-96b3-653642c42b9c-operator-scripts\") pod \"nova-cell1-db-create-bkbz5\" (UID: \"24282d99-b526-4bd3-96b3-653642c42b9c\") " pod="openstack/nova-cell1-db-create-bkbz5" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.941339 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fp892\" (UniqueName: \"kubernetes.io/projected/24282d99-b526-4bd3-96b3-653642c42b9c-kube-api-access-fp892\") pod \"nova-cell1-db-create-bkbz5\" (UID: \"24282d99-b526-4bd3-96b3-653642c42b9c\") " pod="openstack/nova-cell1-db-create-bkbz5" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.942772 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06fe0f9a-dbc9-4078-92ce-07fe06eb4110-operator-scripts\") pod \"nova-api-f91e-account-create-update-twns5\" (UID: \"06fe0f9a-dbc9-4078-92ce-07fe06eb4110\") " pod="openstack/nova-api-f91e-account-create-update-twns5" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.960287 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jphwx\" (UniqueName: \"kubernetes.io/projected/06fe0f9a-dbc9-4078-92ce-07fe06eb4110-kube-api-access-jphwx\") pod \"nova-api-f91e-account-create-update-twns5\" (UID: \"06fe0f9a-dbc9-4078-92ce-07fe06eb4110\") " pod="openstack/nova-api-f91e-account-create-update-twns5" Jan 26 09:24:53 crc kubenswrapper[4872]: I0126 09:24:53.981962 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5935-account-create-update-ftjg4"] Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.033757 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f91e-account-create-update-twns5" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.043953 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fp892\" (UniqueName: \"kubernetes.io/projected/24282d99-b526-4bd3-96b3-653642c42b9c-kube-api-access-fp892\") pod \"nova-cell1-db-create-bkbz5\" (UID: \"24282d99-b526-4bd3-96b3-653642c42b9c\") " pod="openstack/nova-cell1-db-create-bkbz5" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.044264 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d61a34c-59b3-490d-94a2-b9186426f179-operator-scripts\") pod \"nova-cell0-5935-account-create-update-ftjg4\" (UID: \"8d61a34c-59b3-490d-94a2-b9186426f179\") " pod="openstack/nova-cell0-5935-account-create-update-ftjg4" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.044321 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bthcr\" (UniqueName: \"kubernetes.io/projected/8d61a34c-59b3-490d-94a2-b9186426f179-kube-api-access-bthcr\") pod \"nova-cell0-5935-account-create-update-ftjg4\" (UID: \"8d61a34c-59b3-490d-94a2-b9186426f179\") " pod="openstack/nova-cell0-5935-account-create-update-ftjg4" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.044454 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24282d99-b526-4bd3-96b3-653642c42b9c-operator-scripts\") pod \"nova-cell1-db-create-bkbz5\" (UID: \"24282d99-b526-4bd3-96b3-653642c42b9c\") " pod="openstack/nova-cell1-db-create-bkbz5" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.046336 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24282d99-b526-4bd3-96b3-653642c42b9c-operator-scripts\") pod \"nova-cell1-db-create-bkbz5\" (UID: \"24282d99-b526-4bd3-96b3-653642c42b9c\") " pod="openstack/nova-cell1-db-create-bkbz5" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.074679 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fp892\" (UniqueName: \"kubernetes.io/projected/24282d99-b526-4bd3-96b3-653642c42b9c-kube-api-access-fp892\") pod \"nova-cell1-db-create-bkbz5\" (UID: \"24282d99-b526-4bd3-96b3-653642c42b9c\") " pod="openstack/nova-cell1-db-create-bkbz5" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.106779 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-4bea-account-create-update-vh8qs"] Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.111279 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.113836 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.136863 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4bea-account-create-update-vh8qs"] Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.152937 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bthcr\" (UniqueName: \"kubernetes.io/projected/8d61a34c-59b3-490d-94a2-b9186426f179-kube-api-access-bthcr\") pod \"nova-cell0-5935-account-create-update-ftjg4\" (UID: \"8d61a34c-59b3-490d-94a2-b9186426f179\") " pod="openstack/nova-cell0-5935-account-create-update-ftjg4" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.153171 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d61a34c-59b3-490d-94a2-b9186426f179-operator-scripts\") pod \"nova-cell0-5935-account-create-update-ftjg4\" (UID: \"8d61a34c-59b3-490d-94a2-b9186426f179\") " pod="openstack/nova-cell0-5935-account-create-update-ftjg4" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.154404 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d61a34c-59b3-490d-94a2-b9186426f179-operator-scripts\") pod \"nova-cell0-5935-account-create-update-ftjg4\" (UID: \"8d61a34c-59b3-490d-94a2-b9186426f179\") " pod="openstack/nova-cell0-5935-account-create-update-ftjg4" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.154731 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bkbz5" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.189690 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bthcr\" (UniqueName: \"kubernetes.io/projected/8d61a34c-59b3-490d-94a2-b9186426f179-kube-api-access-bthcr\") pod \"nova-cell0-5935-account-create-update-ftjg4\" (UID: \"8d61a34c-59b3-490d-94a2-b9186426f179\") " pod="openstack/nova-cell0-5935-account-create-update-ftjg4" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.245299 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5935-account-create-update-ftjg4" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.257222 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/574bcef8-68d6-40d8-adbb-5a321a3b59ce-operator-scripts\") pod \"nova-cell1-4bea-account-create-update-vh8qs\" (UID: \"574bcef8-68d6-40d8-adbb-5a321a3b59ce\") " pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.257534 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z97fs\" (UniqueName: \"kubernetes.io/projected/574bcef8-68d6-40d8-adbb-5a321a3b59ce-kube-api-access-z97fs\") pod \"nova-cell1-4bea-account-create-update-vh8qs\" (UID: \"574bcef8-68d6-40d8-adbb-5a321a3b59ce\") " pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.360308 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z97fs\" (UniqueName: \"kubernetes.io/projected/574bcef8-68d6-40d8-adbb-5a321a3b59ce-kube-api-access-z97fs\") pod \"nova-cell1-4bea-account-create-update-vh8qs\" (UID: \"574bcef8-68d6-40d8-adbb-5a321a3b59ce\") " pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.360453 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/574bcef8-68d6-40d8-adbb-5a321a3b59ce-operator-scripts\") pod \"nova-cell1-4bea-account-create-update-vh8qs\" (UID: \"574bcef8-68d6-40d8-adbb-5a321a3b59ce\") " pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.361473 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/574bcef8-68d6-40d8-adbb-5a321a3b59ce-operator-scripts\") pod \"nova-cell1-4bea-account-create-update-vh8qs\" (UID: \"574bcef8-68d6-40d8-adbb-5a321a3b59ce\") " pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.378262 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z97fs\" (UniqueName: \"kubernetes.io/projected/574bcef8-68d6-40d8-adbb-5a321a3b59ce-kube-api-access-z97fs\") pod \"nova-cell1-4bea-account-create-update-vh8qs\" (UID: \"574bcef8-68d6-40d8-adbb-5a321a3b59ce\") " pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" Jan 26 09:24:54 crc kubenswrapper[4872]: I0126 09:24:54.463865 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" Jan 26 09:24:55 crc kubenswrapper[4872]: I0126 09:24:55.333630 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:55 crc kubenswrapper[4872]: I0126 09:24:55.350098 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-68bdb4f9bc-lg4kg" Jan 26 09:24:56 crc kubenswrapper[4872]: I0126 09:24:56.007886 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:24:56 crc kubenswrapper[4872]: I0126 09:24:56.008460 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d24cd308-eaee-4bf0-be25-f4822e095b22" containerName="glance-log" containerID="cri-o://8e8de46c5ab125013aadce3938b5a92e6d301d8280ffd9c143b6ebeba3b60d67" gracePeriod=30 Jan 26 09:24:56 crc kubenswrapper[4872]: I0126 09:24:56.008607 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d24cd308-eaee-4bf0-be25-f4822e095b22" containerName="glance-httpd" containerID="cri-o://56b0345c2ca2bf94f06856c8eed362b0931457cd42f4990400a9cf7d0ab5717e" gracePeriod=30 Jan 26 09:24:56 crc kubenswrapper[4872]: I0126 09:24:56.490069 4872 generic.go:334] "Generic (PLEG): container finished" podID="d24cd308-eaee-4bf0-be25-f4822e095b22" containerID="8e8de46c5ab125013aadce3938b5a92e6d301d8280ffd9c143b6ebeba3b60d67" exitCode=143 Jan 26 09:24:56 crc kubenswrapper[4872]: I0126 09:24:56.490125 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d24cd308-eaee-4bf0-be25-f4822e095b22","Type":"ContainerDied","Data":"8e8de46c5ab125013aadce3938b5a92e6d301d8280ffd9c143b6ebeba3b60d67"} Jan 26 09:24:57 crc kubenswrapper[4872]: I0126 09:24:57.211438 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:24:57 crc kubenswrapper[4872]: I0126 09:24:57.212126 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a7856628-f316-416e-a721-bb74270c6a02" containerName="glance-log" containerID="cri-o://d77a7b2f69bed239f43fb42d46d990f2e6b7804e92735447f7dfd1ca1783642e" gracePeriod=30 Jan 26 09:24:57 crc kubenswrapper[4872]: I0126 09:24:57.212328 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a7856628-f316-416e-a721-bb74270c6a02" containerName="glance-httpd" containerID="cri-o://54460e5c70948d40133e61aacdba15fd35cb02c9a163eac703fc6911a158ef3c" gracePeriod=30 Jan 26 09:24:57 crc kubenswrapper[4872]: I0126 09:24:57.501930 4872 generic.go:334] "Generic (PLEG): container finished" podID="a7856628-f316-416e-a721-bb74270c6a02" containerID="d77a7b2f69bed239f43fb42d46d990f2e6b7804e92735447f7dfd1ca1783642e" exitCode=143 Jan 26 09:24:57 crc kubenswrapper[4872]: I0126 09:24:57.501980 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a7856628-f316-416e-a721-bb74270c6a02","Type":"ContainerDied","Data":"d77a7b2f69bed239f43fb42d46d990f2e6b7804e92735447f7dfd1ca1783642e"} Jan 26 09:24:58 crc kubenswrapper[4872]: I0126 09:24:58.184729 4872 scope.go:117] "RemoveContainer" containerID="4e2e1a239f503efca62c57a52ec0ca4f72685e7934b746cee1b12681d3e89b09" Jan 26 09:24:59 crc kubenswrapper[4872]: I0126 09:24:59.170475 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="d24cd308-eaee-4bf0-be25-f4822e095b22" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.149:9292/healthcheck\": read tcp 10.217.0.2:42126->10.217.0.149:9292: read: connection reset by peer" Jan 26 09:24:59 crc kubenswrapper[4872]: I0126 09:24:59.177499 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="d24cd308-eaee-4bf0-be25-f4822e095b22" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.149:9292/healthcheck\": read tcp 10.217.0.2:42132->10.217.0.149:9292: read: connection reset by peer" Jan 26 09:24:59 crc kubenswrapper[4872]: I0126 09:24:59.528496 4872 generic.go:334] "Generic (PLEG): container finished" podID="d24cd308-eaee-4bf0-be25-f4822e095b22" containerID="56b0345c2ca2bf94f06856c8eed362b0931457cd42f4990400a9cf7d0ab5717e" exitCode=0 Jan 26 09:24:59 crc kubenswrapper[4872]: I0126 09:24:59.528554 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d24cd308-eaee-4bf0-be25-f4822e095b22","Type":"ContainerDied","Data":"56b0345c2ca2bf94f06856c8eed362b0931457cd42f4990400a9cf7d0ab5717e"} Jan 26 09:25:00 crc kubenswrapper[4872]: I0126 09:25:00.370900 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="a7856628-f316-416e-a721-bb74270c6a02" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.150:9292/healthcheck\": read tcp 10.217.0.2:34676->10.217.0.150:9292: read: connection reset by peer" Jan 26 09:25:00 crc kubenswrapper[4872]: I0126 09:25:00.370961 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="a7856628-f316-416e-a721-bb74270c6a02" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.150:9292/healthcheck\": read tcp 10.217.0.2:34670->10.217.0.150:9292: read: connection reset by peer" Jan 26 09:25:00 crc kubenswrapper[4872]: I0126 09:25:00.547384 4872 generic.go:334] "Generic (PLEG): container finished" podID="a7856628-f316-416e-a721-bb74270c6a02" containerID="54460e5c70948d40133e61aacdba15fd35cb02c9a163eac703fc6911a158ef3c" exitCode=0 Jan 26 09:25:00 crc kubenswrapper[4872]: I0126 09:25:00.547454 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a7856628-f316-416e-a721-bb74270c6a02","Type":"ContainerDied","Data":"54460e5c70948d40133e61aacdba15fd35cb02c9a163eac703fc6911a158ef3c"} Jan 26 09:25:02 crc kubenswrapper[4872]: E0126 09:25:02.424242 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified" Jan 26 09:25:02 crc kubenswrapper[4872]: E0126 09:25:02.425478 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d7h659h564h565h57chcfh8ch568h664h79h698h684h675h68bh5cdh658h66fh5f6h659h579h59dh694h648h9dh579h566h65bh68bhfch5c9h89h598q,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qwncl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(52a92868-e230-4403-9297-51c2b457343b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:25:02 crc kubenswrapper[4872]: E0126 09:25:02.426991 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="52a92868-e230-4403-9297-51c2b457343b" Jan 26 09:25:02 crc kubenswrapper[4872]: E0126 09:25:02.607750 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\"\"" pod="openstack/openstackclient" podUID="52a92868-e230-4403-9297-51c2b457343b" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.177211 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5935-account-create-update-ftjg4"] Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.182747 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.255059 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d24cd308-eaee-4bf0-be25-f4822e095b22-httpd-run\") pod \"d24cd308-eaee-4bf0-be25-f4822e095b22\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.255134 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cx9m5\" (UniqueName: \"kubernetes.io/projected/d24cd308-eaee-4bf0-be25-f4822e095b22-kube-api-access-cx9m5\") pod \"d24cd308-eaee-4bf0-be25-f4822e095b22\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.255181 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-scripts\") pod \"d24cd308-eaee-4bf0-be25-f4822e095b22\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.255243 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-public-tls-certs\") pod \"d24cd308-eaee-4bf0-be25-f4822e095b22\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.255299 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d24cd308-eaee-4bf0-be25-f4822e095b22-logs\") pod \"d24cd308-eaee-4bf0-be25-f4822e095b22\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.255429 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-combined-ca-bundle\") pod \"d24cd308-eaee-4bf0-be25-f4822e095b22\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.255452 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-config-data\") pod \"d24cd308-eaee-4bf0-be25-f4822e095b22\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.255474 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"d24cd308-eaee-4bf0-be25-f4822e095b22\" (UID: \"d24cd308-eaee-4bf0-be25-f4822e095b22\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.257355 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d24cd308-eaee-4bf0-be25-f4822e095b22-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d24cd308-eaee-4bf0-be25-f4822e095b22" (UID: "d24cd308-eaee-4bf0-be25-f4822e095b22"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.258172 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d24cd308-eaee-4bf0-be25-f4822e095b22-logs" (OuterVolumeSpecName: "logs") pod "d24cd308-eaee-4bf0-be25-f4822e095b22" (UID: "d24cd308-eaee-4bf0-be25-f4822e095b22"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.280486 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-scripts" (OuterVolumeSpecName: "scripts") pod "d24cd308-eaee-4bf0-be25-f4822e095b22" (UID: "d24cd308-eaee-4bf0-be25-f4822e095b22"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.284110 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "d24cd308-eaee-4bf0-be25-f4822e095b22" (UID: "d24cd308-eaee-4bf0-be25-f4822e095b22"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.288245 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d24cd308-eaee-4bf0-be25-f4822e095b22-kube-api-access-cx9m5" (OuterVolumeSpecName: "kube-api-access-cx9m5") pod "d24cd308-eaee-4bf0-be25-f4822e095b22" (UID: "d24cd308-eaee-4bf0-be25-f4822e095b22"). InnerVolumeSpecName "kube-api-access-cx9m5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.317825 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.374630 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"a7856628-f316-416e-a721-bb74270c6a02\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.375841 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-config-data\") pod \"a7856628-f316-416e-a721-bb74270c6a02\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.375933 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-combined-ca-bundle\") pod \"a7856628-f316-416e-a721-bb74270c6a02\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.376052 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7856628-f316-416e-a721-bb74270c6a02-logs\") pod \"a7856628-f316-416e-a721-bb74270c6a02\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.376149 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-internal-tls-certs\") pod \"a7856628-f316-416e-a721-bb74270c6a02\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.376253 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtwq5\" (UniqueName: \"kubernetes.io/projected/a7856628-f316-416e-a721-bb74270c6a02-kube-api-access-vtwq5\") pod \"a7856628-f316-416e-a721-bb74270c6a02\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.376331 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-scripts\") pod \"a7856628-f316-416e-a721-bb74270c6a02\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.377322 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7856628-f316-416e-a721-bb74270c6a02-httpd-run\") pod \"a7856628-f316-416e-a721-bb74270c6a02\" (UID: \"a7856628-f316-416e-a721-bb74270c6a02\") " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.377964 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7856628-f316-416e-a721-bb74270c6a02-logs" (OuterVolumeSpecName: "logs") pod "a7856628-f316-416e-a721-bb74270c6a02" (UID: "a7856628-f316-416e-a721-bb74270c6a02"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.378337 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7856628-f316-416e-a721-bb74270c6a02-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.378430 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.378507 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d24cd308-eaee-4bf0-be25-f4822e095b22-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.378575 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.378691 4872 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d24cd308-eaee-4bf0-be25-f4822e095b22-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.378748 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cx9m5\" (UniqueName: \"kubernetes.io/projected/d24cd308-eaee-4bf0-be25-f4822e095b22-kube-api-access-cx9m5\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.390265 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7856628-f316-416e-a721-bb74270c6a02-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a7856628-f316-416e-a721-bb74270c6a02" (UID: "a7856628-f316-416e-a721-bb74270c6a02"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.413892 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "a7856628-f316-416e-a721-bb74270c6a02" (UID: "a7856628-f316-416e-a721-bb74270c6a02"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.413929 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7856628-f316-416e-a721-bb74270c6a02-kube-api-access-vtwq5" (OuterVolumeSpecName: "kube-api-access-vtwq5") pod "a7856628-f316-416e-a721-bb74270c6a02" (UID: "a7856628-f316-416e-a721-bb74270c6a02"). InnerVolumeSpecName "kube-api-access-vtwq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.414173 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-scripts" (OuterVolumeSpecName: "scripts") pod "a7856628-f316-416e-a721-bb74270c6a02" (UID: "a7856628-f316-416e-a721-bb74270c6a02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.480563 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.486190 4872 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a7856628-f316-416e-a721-bb74270c6a02-httpd-run\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.500858 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.500905 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtwq5\" (UniqueName: \"kubernetes.io/projected/a7856628-f316-416e-a721-bb74270c6a02-kube-api-access-vtwq5\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.570348 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.575566 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d24cd308-eaee-4bf0-be25-f4822e095b22" (UID: "d24cd308-eaee-4bf0-be25-f4822e095b22"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.588997 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7856628-f316-416e-a721-bb74270c6a02" (UID: "a7856628-f316-416e-a721-bb74270c6a02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.604044 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.604136 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.604860 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.604982 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.605060 4872 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.613398 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d24cd308-eaee-4bf0-be25-f4822e095b22" (UID: "d24cd308-eaee-4bf0-be25-f4822e095b22"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.628329 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5935-account-create-update-ftjg4" event={"ID":"8d61a34c-59b3-490d-94a2-b9186426f179","Type":"ContainerStarted","Data":"9a89cae1dc3a745455c3be3665e581763c10ea569d4d8f73104b4ce7d57a667a"} Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.646177 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-config-data" (OuterVolumeSpecName: "config-data") pod "d24cd308-eaee-4bf0-be25-f4822e095b22" (UID: "d24cd308-eaee-4bf0-be25-f4822e095b22"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.674450 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94a8957b-b742-4dfd-a463-e265cf4db32e","Type":"ContainerStarted","Data":"2950135b35b063ebe45e02bfcda7adcaf3a0484a7845bb34ea8eeb9a995655e3"} Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.674664 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-4bea-account-create-update-vh8qs"] Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.685585 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cl2gt"] Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.688220 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" event={"ID":"f721aeee-7daf-413d-885b-5c146881eb99","Type":"ContainerStarted","Data":"4257303280c6aa0818d7a3469f31ab0a87f3c6ef87fb5f5ae944311e72de51b6"} Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.689319 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.697612 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-6dpv7"] Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.707883 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.707916 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d24cd308-eaee-4bf0-be25-f4822e095b22-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.707945 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f91e-account-create-update-twns5"] Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.717898 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-bkbz5"] Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.730784 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.730781 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a7856628-f316-416e-a721-bb74270c6a02","Type":"ContainerDied","Data":"5f94a99ecf19d7abf57d2cd9f49d9d1bfcf1492bbf84aaebbc7e0ea7789d475a"} Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.732219 4872 scope.go:117] "RemoveContainer" containerID="54460e5c70948d40133e61aacdba15fd35cb02c9a163eac703fc6911a158ef3c" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.737139 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-0"] Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.737602 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-config-data" (OuterVolumeSpecName: "config-data") pod "a7856628-f316-416e-a721-bb74270c6a02" (UID: "a7856628-f316-416e-a721-bb74270c6a02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.753311 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a7856628-f316-416e-a721-bb74270c6a02" (UID: "a7856628-f316-416e-a721-bb74270c6a02"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.774407 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d24cd308-eaee-4bf0-be25-f4822e095b22","Type":"ContainerDied","Data":"8e29cbee17b05120d742926e8135c716093985cd80e6b003d187a062ec5475c2"} Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.774751 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.862067 4872 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.862437 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7856628-f316-416e-a721-bb74270c6a02-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.946983 4872 scope.go:117] "RemoveContainer" containerID="d77a7b2f69bed239f43fb42d46d990f2e6b7804e92735447f7dfd1ca1783642e" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.947883 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.972740 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.997022 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.997108 4872 scope.go:117] "RemoveContainer" containerID="56b0345c2ca2bf94f06856c8eed362b0931457cd42f4990400a9cf7d0ab5717e" Jan 26 09:25:03 crc kubenswrapper[4872]: E0126 09:25:03.997672 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d24cd308-eaee-4bf0-be25-f4822e095b22" containerName="glance-log" Jan 26 09:25:03 crc kubenswrapper[4872]: I0126 09:25:03.997696 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d24cd308-eaee-4bf0-be25-f4822e095b22" containerName="glance-log" Jan 26 09:25:04 crc kubenswrapper[4872]: E0126 09:25:03.998262 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7856628-f316-416e-a721-bb74270c6a02" containerName="glance-httpd" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:03.998281 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7856628-f316-416e-a721-bb74270c6a02" containerName="glance-httpd" Jan 26 09:25:04 crc kubenswrapper[4872]: E0126 09:25:03.998290 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7856628-f316-416e-a721-bb74270c6a02" containerName="glance-log" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:03.998298 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7856628-f316-416e-a721-bb74270c6a02" containerName="glance-log" Jan 26 09:25:04 crc kubenswrapper[4872]: E0126 09:25:03.998341 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d24cd308-eaee-4bf0-be25-f4822e095b22" containerName="glance-httpd" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:03.998350 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d24cd308-eaee-4bf0-be25-f4822e095b22" containerName="glance-httpd" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:03.998599 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d24cd308-eaee-4bf0-be25-f4822e095b22" containerName="glance-log" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:03.998623 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7856628-f316-416e-a721-bb74270c6a02" containerName="glance-log" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:03.998632 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d24cd308-eaee-4bf0-be25-f4822e095b22" containerName="glance-httpd" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:03.998646 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7856628-f316-416e-a721-bb74270c6a02" containerName="glance-httpd" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.000411 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.016960 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.017067 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.017107 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.067253 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e79d158-7664-4bfd-98b8-8cbf9092584a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.067343 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smp56\" (UniqueName: \"kubernetes.io/projected/4e79d158-7664-4bfd-98b8-8cbf9092584a-kube-api-access-smp56\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.067392 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e79d158-7664-4bfd-98b8-8cbf9092584a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.067414 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e79d158-7664-4bfd-98b8-8cbf9092584a-scripts\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.067477 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.067520 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4e79d158-7664-4bfd-98b8-8cbf9092584a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.067547 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e79d158-7664-4bfd-98b8-8cbf9092584a-logs\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.067607 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e79d158-7664-4bfd-98b8-8cbf9092584a-config-data\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.123081 4872 scope.go:117] "RemoveContainer" containerID="8e8de46c5ab125013aadce3938b5a92e6d301d8280ffd9c143b6ebeba3b60d67" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.127186 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.150010 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.170092 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4e79d158-7664-4bfd-98b8-8cbf9092584a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.170485 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e79d158-7664-4bfd-98b8-8cbf9092584a-logs\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.170547 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e79d158-7664-4bfd-98b8-8cbf9092584a-config-data\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.170573 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e79d158-7664-4bfd-98b8-8cbf9092584a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.170632 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smp56\" (UniqueName: \"kubernetes.io/projected/4e79d158-7664-4bfd-98b8-8cbf9092584a-kube-api-access-smp56\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.170679 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e79d158-7664-4bfd-98b8-8cbf9092584a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.170701 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e79d158-7664-4bfd-98b8-8cbf9092584a-scripts\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.170763 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.171256 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.180536 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4e79d158-7664-4bfd-98b8-8cbf9092584a-logs\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.180773 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4e79d158-7664-4bfd-98b8-8cbf9092584a-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.192587 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.194717 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.197375 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4e79d158-7664-4bfd-98b8-8cbf9092584a-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.204552 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e79d158-7664-4bfd-98b8-8cbf9092584a-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.205088 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.205185 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.208779 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.212871 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smp56\" (UniqueName: \"kubernetes.io/projected/4e79d158-7664-4bfd-98b8-8cbf9092584a-kube-api-access-smp56\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.213225 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e79d158-7664-4bfd-98b8-8cbf9092584a-config-data\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.213403 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4e79d158-7664-4bfd-98b8-8cbf9092584a-scripts\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.286253 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.286357 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.286448 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.286518 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.286541 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.286558 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.286609 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdf7m\" (UniqueName: \"kubernetes.io/projected/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-kube-api-access-wdf7m\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.286663 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.343880 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"4e79d158-7664-4bfd-98b8-8cbf9092584a\") " pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.388184 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.388249 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.388285 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.388320 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.388339 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.388359 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.388381 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdf7m\" (UniqueName: \"kubernetes.io/projected/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-kube-api-access-wdf7m\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.388407 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.389053 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.392520 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-logs\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.392988 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.393905 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.397044 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.399199 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.403549 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.415574 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdf7m\" (UniqueName: \"kubernetes.io/projected/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-kube-api-access-wdf7m\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.416467 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.436400 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6\") " pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.531416 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.789489 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6dpv7" event={"ID":"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff","Type":"ContainerStarted","Data":"ee02f06a56b72178f0d9662be06e1101af2d56c208359c4dd1c813b29bdd473e"} Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.789999 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6dpv7" event={"ID":"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff","Type":"ContainerStarted","Data":"2f64f9c874ba7dcdaf70ad6327004acdd941abc3db4278d2b75a7d0d38b160d1"} Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.796778 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cl2gt" event={"ID":"4f99a167-975f-4373-8c9d-23a2ab6510fa","Type":"ContainerStarted","Data":"b057a2f81ec82a726a1722537c79e5488dd003bd9df3b4017ede5e83a731656e"} Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.796855 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cl2gt" event={"ID":"4f99a167-975f-4373-8c9d-23a2ab6510fa","Type":"ContainerStarted","Data":"3f148081de1320ef007a1a85d0f75ef3527be5596d7d193ab8b398c1ac52f5f3"} Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.799118 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"f8a676a6-3108-4171-82c0-9f7c08e521f8","Type":"ContainerStarted","Data":"bdfea61b08735188baf1427f52cae352787c91e1493a62dc6adffd096b3f3178"} Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.806580 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-6dpv7" podStartSLOduration=11.806554804 podStartE2EDuration="11.806554804s" podCreationTimestamp="2026-01-26 09:24:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:25:04.805768944 +0000 UTC m=+1038.114608755" watchObservedRunningTime="2026-01-26 09:25:04.806554804 +0000 UTC m=+1038.115394605" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.808848 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f91e-account-create-update-twns5" event={"ID":"06fe0f9a-dbc9-4078-92ce-07fe06eb4110","Type":"ContainerStarted","Data":"f829bcb22386227b578ea3bcc7d4d268f4fde0ea4a55e5965948a8f04318bafb"} Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.808912 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f91e-account-create-update-twns5" event={"ID":"06fe0f9a-dbc9-4078-92ce-07fe06eb4110","Type":"ContainerStarted","Data":"0a687d34164b4030509586b3f1e2bb5b736df064ac86599ab1dbd0f612a460fb"} Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.823426 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" event={"ID":"574bcef8-68d6-40d8-adbb-5a321a3b59ce","Type":"ContainerStarted","Data":"7c32901cada88266eae0b24839df8e2f9fbbb73d60edebc33f736b085cb61ecf"} Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.823483 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" event={"ID":"574bcef8-68d6-40d8-adbb-5a321a3b59ce","Type":"ContainerStarted","Data":"28eee6f1ac21524e779afac4591e421abeca4bf41b37fe694cbd8db017258745"} Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.831261 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-f91e-account-create-update-twns5" podStartSLOduration=11.831233037 podStartE2EDuration="11.831233037s" podCreationTimestamp="2026-01-26 09:24:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:25:04.826946197 +0000 UTC m=+1038.135786008" watchObservedRunningTime="2026-01-26 09:25:04.831233037 +0000 UTC m=+1038.140072838" Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.838790 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"e135958c-61c2-4199-94b3-2a8e6623310c","Type":"ContainerStarted","Data":"774bcef0a5791129e7b4fdfa77bc83d41208d4e79678cf8d4f5b2373b6f16bc4"} Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.850484 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bkbz5" event={"ID":"24282d99-b526-4bd3-96b3-653642c42b9c","Type":"ContainerStarted","Data":"c55680258422d2aa3b674b16796ad6a3f6e39924b283c61962d5e34b4eeb6390"} Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.850547 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bkbz5" event={"ID":"24282d99-b526-4bd3-96b3-653642c42b9c","Type":"ContainerStarted","Data":"2042141885e0546f9a83958fe7bd0e0a32ba6c3249816b114b1c27fc52d2bed4"} Jan 26 09:25:04 crc kubenswrapper[4872]: I0126 09:25:04.851224 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" podStartSLOduration=10.851203461 podStartE2EDuration="10.851203461s" podCreationTimestamp="2026-01-26 09:24:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:25:04.849355923 +0000 UTC m=+1038.158195724" watchObservedRunningTime="2026-01-26 09:25:04.851203461 +0000 UTC m=+1038.160043282" Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.173154 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.217647 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7856628-f316-416e-a721-bb74270c6a02" path="/var/lib/kubelet/pods/a7856628-f316-416e-a721-bb74270c6a02/volumes" Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.219582 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d24cd308-eaee-4bf0-be25-f4822e095b22" path="/var/lib/kubelet/pods/d24cd308-eaee-4bf0-be25-f4822e095b22/volumes" Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.315036 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Jan 26 09:25:05 crc kubenswrapper[4872]: W0126 09:25:05.319887 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7c8b404_f2e3_4d1e_a5ac_cd6651d48bf6.slice/crio-43d7b4109d65a96ebeb50417300354aea66cf4710d2dc8f8c223f4f411234c7b WatchSource:0}: Error finding container 43d7b4109d65a96ebeb50417300354aea66cf4710d2dc8f8c223f4f411234c7b: Status 404 returned error can't find the container with id 43d7b4109d65a96ebeb50417300354aea66cf4710d2dc8f8c223f4f411234c7b Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.867743 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5935-account-create-update-ftjg4" event={"ID":"8d61a34c-59b3-490d-94a2-b9186426f179","Type":"ContainerStarted","Data":"7b91ffb556bd6ff0114ba45425a3e34c8ebd6365489b8787ecaca6cbf570da49"} Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.869280 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4e79d158-7664-4bfd-98b8-8cbf9092584a","Type":"ContainerStarted","Data":"13f5643a8dede9a40d35cb0b171c6b8643b6fc22a1069ee7b69087f3c9d2dfbb"} Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.871330 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6","Type":"ContainerStarted","Data":"43d7b4109d65a96ebeb50417300354aea66cf4710d2dc8f8c223f4f411234c7b"} Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.881087 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="ceilometer-central-agent" containerID="cri-o://2272551f1062c8f5581604bfa484ed493b479d73101aa07319ce08c2141898ee" gracePeriod=30 Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.882382 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="proxy-httpd" containerID="cri-o://3c6e9a0f17a77512158f836fd07c3ed0b7809f2a23434fd3161ed21bd50540cb" gracePeriod=30 Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.882450 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="sg-core" containerID="cri-o://2950135b35b063ebe45e02bfcda7adcaf3a0484a7845bb34ea8eeb9a995655e3" gracePeriod=30 Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.882489 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="ceilometer-notification-agent" containerID="cri-o://f850eae1ba4ede05658e235af04af5e9ddc091ac135bdddf5732952720385fd6" gracePeriod=30 Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.882686 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.890620 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-5935-account-create-update-ftjg4" podStartSLOduration=12.890596067 podStartE2EDuration="12.890596067s" podCreationTimestamp="2026-01-26 09:24:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:25:05.886329267 +0000 UTC m=+1039.195169078" watchObservedRunningTime="2026-01-26 09:25:05.890596067 +0000 UTC m=+1039.199435868" Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.916026 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-cl2gt" podStartSLOduration=12.916008629 podStartE2EDuration="12.916008629s" podCreationTimestamp="2026-01-26 09:24:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:25:05.904194866 +0000 UTC m=+1039.213034657" watchObservedRunningTime="2026-01-26 09:25:05.916008629 +0000 UTC m=+1039.224848430" Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.945329 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.42983523 podStartE2EDuration="19.945305362s" podCreationTimestamp="2026-01-26 09:24:46 +0000 UTC" firstStartedPulling="2026-01-26 09:24:47.757376814 +0000 UTC m=+1021.066216615" lastFinishedPulling="2026-01-26 09:25:05.272846946 +0000 UTC m=+1038.581686747" observedRunningTime="2026-01-26 09:25:05.937333127 +0000 UTC m=+1039.246172928" watchObservedRunningTime="2026-01-26 09:25:05.945305362 +0000 UTC m=+1039.254145163" Jan 26 09:25:05 crc kubenswrapper[4872]: I0126 09:25:05.961388 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-bkbz5" podStartSLOduration=12.961361284 podStartE2EDuration="12.961361284s" podCreationTimestamp="2026-01-26 09:24:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:25:05.952210739 +0000 UTC m=+1039.261050540" watchObservedRunningTime="2026-01-26 09:25:05.961361284 +0000 UTC m=+1039.270201085" Jan 26 09:25:06 crc kubenswrapper[4872]: I0126 09:25:06.892599 4872 generic.go:334] "Generic (PLEG): container finished" podID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerID="f850eae1ba4ede05658e235af04af5e9ddc091ac135bdddf5732952720385fd6" exitCode=0 Jan 26 09:25:06 crc kubenswrapper[4872]: I0126 09:25:06.892878 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94a8957b-b742-4dfd-a463-e265cf4db32e","Type":"ContainerStarted","Data":"3c6e9a0f17a77512158f836fd07c3ed0b7809f2a23434fd3161ed21bd50540cb"} Jan 26 09:25:06 crc kubenswrapper[4872]: I0126 09:25:06.893157 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94a8957b-b742-4dfd-a463-e265cf4db32e","Type":"ContainerDied","Data":"f850eae1ba4ede05658e235af04af5e9ddc091ac135bdddf5732952720385fd6"} Jan 26 09:25:06 crc kubenswrapper[4872]: I0126 09:25:06.898318 4872 generic.go:334] "Generic (PLEG): container finished" podID="f8a676a6-3108-4171-82c0-9f7c08e521f8" containerID="0f830857927ba81e4498dcb22657f89134c07b4cb2e0689782ed79e3e2947b94" exitCode=0 Jan 26 09:25:06 crc kubenswrapper[4872]: I0126 09:25:06.898417 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"f8a676a6-3108-4171-82c0-9f7c08e521f8","Type":"ContainerDied","Data":"0f830857927ba81e4498dcb22657f89134c07b4cb2e0689782ed79e3e2947b94"} Jan 26 09:25:06 crc kubenswrapper[4872]: I0126 09:25:06.901749 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4e79d158-7664-4bfd-98b8-8cbf9092584a","Type":"ContainerStarted","Data":"805d9268116d20245a5e08a531bc29c69c2d31740e4bf7390cdb04f13acf2d30"} Jan 26 09:25:06 crc kubenswrapper[4872]: I0126 09:25:06.901826 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"4e79d158-7664-4bfd-98b8-8cbf9092584a","Type":"ContainerStarted","Data":"35a8f607238182cb397486f75aacade422621bb25dbd1f3974227e4230886d7a"} Jan 26 09:25:06 crc kubenswrapper[4872]: I0126 09:25:06.904835 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6","Type":"ContainerStarted","Data":"48a56bf6457376059976f14299249da649c292dcb48c306cdbc14e5fa2da78b9"} Jan 26 09:25:06 crc kubenswrapper[4872]: I0126 09:25:06.934038 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.934013137 podStartE2EDuration="3.934013137s" podCreationTimestamp="2026-01-26 09:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:25:06.926309149 +0000 UTC m=+1040.235148950" watchObservedRunningTime="2026-01-26 09:25:06.934013137 +0000 UTC m=+1040.242852938" Jan 26 09:25:07 crc kubenswrapper[4872]: I0126 09:25:07.119581 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:25:07 crc kubenswrapper[4872]: I0126 09:25:07.930953 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6","Type":"ContainerStarted","Data":"f71af9bf4c68c14f1f34e884768c2e9a4a5b1526eeca4b9eaa954ef984159400"} Jan 26 09:25:07 crc kubenswrapper[4872]: I0126 09:25:07.935670 4872 generic.go:334] "Generic (PLEG): container finished" podID="168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff" containerID="ee02f06a56b72178f0d9662be06e1101af2d56c208359c4dd1c813b29bdd473e" exitCode=0 Jan 26 09:25:07 crc kubenswrapper[4872]: I0126 09:25:07.935737 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6dpv7" event={"ID":"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff","Type":"ContainerDied","Data":"ee02f06a56b72178f0d9662be06e1101af2d56c208359c4dd1c813b29bdd473e"} Jan 26 09:25:07 crc kubenswrapper[4872]: I0126 09:25:07.940031 4872 generic.go:334] "Generic (PLEG): container finished" podID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerID="2950135b35b063ebe45e02bfcda7adcaf3a0484a7845bb34ea8eeb9a995655e3" exitCode=2 Jan 26 09:25:07 crc kubenswrapper[4872]: I0126 09:25:07.940070 4872 generic.go:334] "Generic (PLEG): container finished" podID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerID="2272551f1062c8f5581604bfa484ed493b479d73101aa07319ce08c2141898ee" exitCode=0 Jan 26 09:25:07 crc kubenswrapper[4872]: I0126 09:25:07.940113 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94a8957b-b742-4dfd-a463-e265cf4db32e","Type":"ContainerDied","Data":"2950135b35b063ebe45e02bfcda7adcaf3a0484a7845bb34ea8eeb9a995655e3"} Jan 26 09:25:07 crc kubenswrapper[4872]: I0126 09:25:07.940136 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94a8957b-b742-4dfd-a463-e265cf4db32e","Type":"ContainerDied","Data":"2272551f1062c8f5581604bfa484ed493b479d73101aa07319ce08c2141898ee"} Jan 26 09:25:07 crc kubenswrapper[4872]: I0126 09:25:07.942894 4872 generic.go:334] "Generic (PLEG): container finished" podID="e135958c-61c2-4199-94b3-2a8e6623310c" containerID="774bcef0a5791129e7b4fdfa77bc83d41208d4e79678cf8d4f5b2373b6f16bc4" exitCode=0 Jan 26 09:25:07 crc kubenswrapper[4872]: I0126 09:25:07.944203 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"e135958c-61c2-4199-94b3-2a8e6623310c","Type":"ContainerDied","Data":"774bcef0a5791129e7b4fdfa77bc83d41208d4e79678cf8d4f5b2373b6f16bc4"} Jan 26 09:25:07 crc kubenswrapper[4872]: I0126 09:25:07.977816 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.977779056 podStartE2EDuration="3.977779056s" podCreationTimestamp="2026-01-26 09:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:25:07.963709005 +0000 UTC m=+1041.272548816" watchObservedRunningTime="2026-01-26 09:25:07.977779056 +0000 UTC m=+1041.286618857" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.406910 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.513008 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-combined-ca-bundle\") pod \"f8a676a6-3108-4171-82c0-9f7c08e521f8\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.513509 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nff8\" (UniqueName: \"kubernetes.io/projected/f8a676a6-3108-4171-82c0-9f7c08e521f8-kube-api-access-2nff8\") pod \"f8a676a6-3108-4171-82c0-9f7c08e521f8\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.513551 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f8a676a6-3108-4171-82c0-9f7c08e521f8-etc-podinfo\") pod \"f8a676a6-3108-4171-82c0-9f7c08e521f8\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.513596 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-config\") pod \"f8a676a6-3108-4171-82c0-9f7c08e521f8\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.513725 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/f8a676a6-3108-4171-82c0-9f7c08e521f8-var-lib-ironic\") pod \"f8a676a6-3108-4171-82c0-9f7c08e521f8\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.513756 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-scripts\") pod \"f8a676a6-3108-4171-82c0-9f7c08e521f8\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.513877 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/f8a676a6-3108-4171-82c0-9f7c08e521f8-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"f8a676a6-3108-4171-82c0-9f7c08e521f8\" (UID: \"f8a676a6-3108-4171-82c0-9f7c08e521f8\") " Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.514546 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8a676a6-3108-4171-82c0-9f7c08e521f8-var-lib-ironic-inspector-dhcp-hostsdir" (OuterVolumeSpecName: "var-lib-ironic-inspector-dhcp-hostsdir") pod "f8a676a6-3108-4171-82c0-9f7c08e521f8" (UID: "f8a676a6-3108-4171-82c0-9f7c08e521f8"). InnerVolumeSpecName "var-lib-ironic-inspector-dhcp-hostsdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.514574 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8a676a6-3108-4171-82c0-9f7c08e521f8-var-lib-ironic" (OuterVolumeSpecName: "var-lib-ironic") pod "f8a676a6-3108-4171-82c0-9f7c08e521f8" (UID: "f8a676a6-3108-4171-82c0-9f7c08e521f8"). InnerVolumeSpecName "var-lib-ironic". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.522381 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-config" (OuterVolumeSpecName: "config") pod "f8a676a6-3108-4171-82c0-9f7c08e521f8" (UID: "f8a676a6-3108-4171-82c0-9f7c08e521f8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.522403 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8a676a6-3108-4171-82c0-9f7c08e521f8-kube-api-access-2nff8" (OuterVolumeSpecName: "kube-api-access-2nff8") pod "f8a676a6-3108-4171-82c0-9f7c08e521f8" (UID: "f8a676a6-3108-4171-82c0-9f7c08e521f8"). InnerVolumeSpecName "kube-api-access-2nff8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.529141 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-scripts" (OuterVolumeSpecName: "scripts") pod "f8a676a6-3108-4171-82c0-9f7c08e521f8" (UID: "f8a676a6-3108-4171-82c0-9f7c08e521f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.533660 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f8a676a6-3108-4171-82c0-9f7c08e521f8-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "f8a676a6-3108-4171-82c0-9f7c08e521f8" (UID: "f8a676a6-3108-4171-82c0-9f7c08e521f8"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.568088 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8a676a6-3108-4171-82c0-9f7c08e521f8" (UID: "f8a676a6-3108-4171-82c0-9f7c08e521f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.615777 4872 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/f8a676a6-3108-4171-82c0-9f7c08e521f8-var-lib-ironic-inspector-dhcp-hostsdir\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.615832 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.615843 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nff8\" (UniqueName: \"kubernetes.io/projected/f8a676a6-3108-4171-82c0-9f7c08e521f8-kube-api-access-2nff8\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.615853 4872 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f8a676a6-3108-4171-82c0-9f7c08e521f8-etc-podinfo\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.615862 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.615870 4872 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/f8a676a6-3108-4171-82c0-9f7c08e521f8-var-lib-ironic\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.615879 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8a676a6-3108-4171-82c0-9f7c08e521f8-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.970525 4872 generic.go:334] "Generic (PLEG): container finished" podID="06fe0f9a-dbc9-4078-92ce-07fe06eb4110" containerID="f829bcb22386227b578ea3bcc7d4d268f4fde0ea4a55e5965948a8f04318bafb" exitCode=0 Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.970579 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f91e-account-create-update-twns5" event={"ID":"06fe0f9a-dbc9-4078-92ce-07fe06eb4110","Type":"ContainerDied","Data":"f829bcb22386227b578ea3bcc7d4d268f4fde0ea4a55e5965948a8f04318bafb"} Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.972950 4872 generic.go:334] "Generic (PLEG): container finished" podID="574bcef8-68d6-40d8-adbb-5a321a3b59ce" containerID="7c32901cada88266eae0b24839df8e2f9fbbb73d60edebc33f736b085cb61ecf" exitCode=0 Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.973016 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" event={"ID":"574bcef8-68d6-40d8-adbb-5a321a3b59ce","Type":"ContainerDied","Data":"7c32901cada88266eae0b24839df8e2f9fbbb73d60edebc33f736b085cb61ecf"} Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.975084 4872 generic.go:334] "Generic (PLEG): container finished" podID="4f99a167-975f-4373-8c9d-23a2ab6510fa" containerID="b057a2f81ec82a726a1722537c79e5488dd003bd9df3b4017ede5e83a731656e" exitCode=0 Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.975106 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cl2gt" event={"ID":"4f99a167-975f-4373-8c9d-23a2ab6510fa","Type":"ContainerDied","Data":"b057a2f81ec82a726a1722537c79e5488dd003bd9df3b4017ede5e83a731656e"} Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.976884 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"f8a676a6-3108-4171-82c0-9f7c08e521f8","Type":"ContainerDied","Data":"bdfea61b08735188baf1427f52cae352787c91e1493a62dc6adffd096b3f3178"} Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.976946 4872 scope.go:117] "RemoveContainer" containerID="0f830857927ba81e4498dcb22657f89134c07b4cb2e0689782ed79e3e2947b94" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.976944 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.980116 4872 generic.go:334] "Generic (PLEG): container finished" podID="8d61a34c-59b3-490d-94a2-b9186426f179" containerID="7b91ffb556bd6ff0114ba45425a3e34c8ebd6365489b8787ecaca6cbf570da49" exitCode=0 Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.980162 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5935-account-create-update-ftjg4" event={"ID":"8d61a34c-59b3-490d-94a2-b9186426f179","Type":"ContainerDied","Data":"7b91ffb556bd6ff0114ba45425a3e34c8ebd6365489b8787ecaca6cbf570da49"} Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.981971 4872 generic.go:334] "Generic (PLEG): container finished" podID="f721aeee-7daf-413d-885b-5c146881eb99" containerID="4257303280c6aa0818d7a3469f31ab0a87f3c6ef87fb5f5ae944311e72de51b6" exitCode=1 Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.982019 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" event={"ID":"f721aeee-7daf-413d-885b-5c146881eb99","Type":"ContainerDied","Data":"4257303280c6aa0818d7a3469f31ab0a87f3c6ef87fb5f5ae944311e72de51b6"} Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.983219 4872 scope.go:117] "RemoveContainer" containerID="4257303280c6aa0818d7a3469f31ab0a87f3c6ef87fb5f5ae944311e72de51b6" Jan 26 09:25:08 crc kubenswrapper[4872]: E0126 09:25:08.984746 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-5b47d6fc7b-7wkw2_openstack(f721aeee-7daf-413d-885b-5c146881eb99)\"" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" podUID="f721aeee-7daf-413d-885b-5c146881eb99" Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.996955 4872 generic.go:334] "Generic (PLEG): container finished" podID="24282d99-b526-4bd3-96b3-653642c42b9c" containerID="c55680258422d2aa3b674b16796ad6a3f6e39924b283c61962d5e34b4eeb6390" exitCode=0 Jan 26 09:25:08 crc kubenswrapper[4872]: I0126 09:25:08.997070 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bkbz5" event={"ID":"24282d99-b526-4bd3-96b3-653642c42b9c","Type":"ContainerDied","Data":"c55680258422d2aa3b674b16796ad6a3f6e39924b283c61962d5e34b4eeb6390"} Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.130199 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-0"] Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.139552 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-0"] Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.170773 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-0"] Jan 26 09:25:09 crc kubenswrapper[4872]: E0126 09:25:09.171256 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8a676a6-3108-4171-82c0-9f7c08e521f8" containerName="ironic-python-agent-init" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.171270 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8a676a6-3108-4171-82c0-9f7c08e521f8" containerName="ironic-python-agent-init" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.171452 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8a676a6-3108-4171-82c0-9f7c08e521f8" containerName="ironic-python-agent-init" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.188118 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.193083 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-inspector-public-svc" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.193369 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-inspector-internal-svc" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.193658 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.193687 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.241454 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8a676a6-3108-4171-82c0-9f7c08e521f8" path="/var/lib/kubelet/pods/f8a676a6-3108-4171-82c0-9f7c08e521f8/volumes" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.290813 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.345505 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.345672 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.345736 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/8865ee9d-9129-488e-84e2-cf1a11547f8a-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.345850 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-scripts\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.346092 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.346146 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq7cx\" (UniqueName: \"kubernetes.io/projected/8865ee9d-9129-488e-84e2-cf1a11547f8a-kube-api-access-lq7cx\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.346189 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/8865ee9d-9129-488e-84e2-cf1a11547f8a-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.346269 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-config\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.346293 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/8865ee9d-9129-488e-84e2-cf1a11547f8a-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.448761 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.448861 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/8865ee9d-9129-488e-84e2-cf1a11547f8a-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.448949 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-scripts\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.448995 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.449019 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq7cx\" (UniqueName: \"kubernetes.io/projected/8865ee9d-9129-488e-84e2-cf1a11547f8a-kube-api-access-lq7cx\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.449043 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/8865ee9d-9129-488e-84e2-cf1a11547f8a-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.449081 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-config\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.449098 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/8865ee9d-9129-488e-84e2-cf1a11547f8a-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.449143 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.450922 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/8865ee9d-9129-488e-84e2-cf1a11547f8a-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.451099 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/8865ee9d-9129-488e-84e2-cf1a11547f8a-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.456498 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.457035 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/8865ee9d-9129-488e-84e2-cf1a11547f8a-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.457614 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.458944 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-config\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.459258 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.468929 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8865ee9d-9129-488e-84e2-cf1a11547f8a-scripts\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.472009 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq7cx\" (UniqueName: \"kubernetes.io/projected/8865ee9d-9129-488e-84e2-cf1a11547f8a-kube-api-access-lq7cx\") pod \"ironic-inspector-0\" (UID: \"8865ee9d-9129-488e-84e2-cf1a11547f8a\") " pod="openstack/ironic-inspector-0" Jan 26 09:25:09 crc kubenswrapper[4872]: I0126 09:25:09.588011 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Jan 26 09:25:10 crc kubenswrapper[4872]: I0126 09:25:10.837333 4872 scope.go:117] "RemoveContainer" containerID="4e2e1a239f503efca62c57a52ec0ca4f72685e7934b746cee1b12681d3e89b09" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.037738 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cl2gt" event={"ID":"4f99a167-975f-4373-8c9d-23a2ab6510fa","Type":"ContainerDied","Data":"3f148081de1320ef007a1a85d0f75ef3527be5596d7d193ab8b398c1ac52f5f3"} Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.037787 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f148081de1320ef007a1a85d0f75ef3527be5596d7d193ab8b398c1ac52f5f3" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.039863 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6dpv7" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.052685 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5935-account-create-update-ftjg4" event={"ID":"8d61a34c-59b3-490d-94a2-b9186426f179","Type":"ContainerDied","Data":"9a89cae1dc3a745455c3be3665e581763c10ea569d4d8f73104b4ce7d57a667a"} Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.052729 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a89cae1dc3a745455c3be3665e581763c10ea569d4d8f73104b4ce7d57a667a" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.053119 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5935-account-create-update-ftjg4" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.077826 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cl2gt" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.079927 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bkbz5" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.078046 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-bkbz5" event={"ID":"24282d99-b526-4bd3-96b3-653642c42b9c","Type":"ContainerDied","Data":"2042141885e0546f9a83958fe7bd0e0a32ba6c3249816b114b1c27fc52d2bed4"} Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.080103 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2042141885e0546f9a83958fe7bd0e0a32ba6c3249816b114b1c27fc52d2bed4" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.082098 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f91e-account-create-update-twns5" event={"ID":"06fe0f9a-dbc9-4078-92ce-07fe06eb4110","Type":"ContainerDied","Data":"0a687d34164b4030509586b3f1e2bb5b736df064ac86599ab1dbd0f612a460fb"} Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.082145 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a687d34164b4030509586b3f1e2bb5b736df064ac86599ab1dbd0f612a460fb" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.084340 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-6dpv7" event={"ID":"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff","Type":"ContainerDied","Data":"2f64f9c874ba7dcdaf70ad6327004acdd941abc3db4278d2b75a7d0d38b160d1"} Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.084371 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f64f9c874ba7dcdaf70ad6327004acdd941abc3db4278d2b75a7d0d38b160d1" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.084462 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-6dpv7" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.087641 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f91e-account-create-update-twns5" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.088361 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.088509 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" event={"ID":"574bcef8-68d6-40d8-adbb-5a321a3b59ce","Type":"ContainerDied","Data":"28eee6f1ac21524e779afac4591e421abeca4bf41b37fe694cbd8db017258745"} Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.088543 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28eee6f1ac21524e779afac4591e421abeca4bf41b37fe694cbd8db017258745" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.116279 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jphwx\" (UniqueName: \"kubernetes.io/projected/06fe0f9a-dbc9-4078-92ce-07fe06eb4110-kube-api-access-jphwx\") pod \"06fe0f9a-dbc9-4078-92ce-07fe06eb4110\" (UID: \"06fe0f9a-dbc9-4078-92ce-07fe06eb4110\") " Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.116365 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fp892\" (UniqueName: \"kubernetes.io/projected/24282d99-b526-4bd3-96b3-653642c42b9c-kube-api-access-fp892\") pod \"24282d99-b526-4bd3-96b3-653642c42b9c\" (UID: \"24282d99-b526-4bd3-96b3-653642c42b9c\") " Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.116400 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff-operator-scripts\") pod \"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff\" (UID: \"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff\") " Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.116430 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgb9k\" (UniqueName: \"kubernetes.io/projected/4f99a167-975f-4373-8c9d-23a2ab6510fa-kube-api-access-fgb9k\") pod \"4f99a167-975f-4373-8c9d-23a2ab6510fa\" (UID: \"4f99a167-975f-4373-8c9d-23a2ab6510fa\") " Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.116511 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24282d99-b526-4bd3-96b3-653642c42b9c-operator-scripts\") pod \"24282d99-b526-4bd3-96b3-653642c42b9c\" (UID: \"24282d99-b526-4bd3-96b3-653642c42b9c\") " Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.116539 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2sl5x\" (UniqueName: \"kubernetes.io/projected/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff-kube-api-access-2sl5x\") pod \"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff\" (UID: \"168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff\") " Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.116563 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/574bcef8-68d6-40d8-adbb-5a321a3b59ce-operator-scripts\") pod \"574bcef8-68d6-40d8-adbb-5a321a3b59ce\" (UID: \"574bcef8-68d6-40d8-adbb-5a321a3b59ce\") " Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.116596 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z97fs\" (UniqueName: \"kubernetes.io/projected/574bcef8-68d6-40d8-adbb-5a321a3b59ce-kube-api-access-z97fs\") pod \"574bcef8-68d6-40d8-adbb-5a321a3b59ce\" (UID: \"574bcef8-68d6-40d8-adbb-5a321a3b59ce\") " Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.116659 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d61a34c-59b3-490d-94a2-b9186426f179-operator-scripts\") pod \"8d61a34c-59b3-490d-94a2-b9186426f179\" (UID: \"8d61a34c-59b3-490d-94a2-b9186426f179\") " Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.116681 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06fe0f9a-dbc9-4078-92ce-07fe06eb4110-operator-scripts\") pod \"06fe0f9a-dbc9-4078-92ce-07fe06eb4110\" (UID: \"06fe0f9a-dbc9-4078-92ce-07fe06eb4110\") " Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.116703 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f99a167-975f-4373-8c9d-23a2ab6510fa-operator-scripts\") pod \"4f99a167-975f-4373-8c9d-23a2ab6510fa\" (UID: \"4f99a167-975f-4373-8c9d-23a2ab6510fa\") " Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.116719 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bthcr\" (UniqueName: \"kubernetes.io/projected/8d61a34c-59b3-490d-94a2-b9186426f179-kube-api-access-bthcr\") pod \"8d61a34c-59b3-490d-94a2-b9186426f179\" (UID: \"8d61a34c-59b3-490d-94a2-b9186426f179\") " Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.120182 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d61a34c-59b3-490d-94a2-b9186426f179-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8d61a34c-59b3-490d-94a2-b9186426f179" (UID: "8d61a34c-59b3-490d-94a2-b9186426f179"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.120241 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff" (UID: "168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.120719 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/574bcef8-68d6-40d8-adbb-5a321a3b59ce-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "574bcef8-68d6-40d8-adbb-5a321a3b59ce" (UID: "574bcef8-68d6-40d8-adbb-5a321a3b59ce"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.123322 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06fe0f9a-dbc9-4078-92ce-07fe06eb4110-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "06fe0f9a-dbc9-4078-92ce-07fe06eb4110" (UID: "06fe0f9a-dbc9-4078-92ce-07fe06eb4110"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.123590 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4f99a167-975f-4373-8c9d-23a2ab6510fa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4f99a167-975f-4373-8c9d-23a2ab6510fa" (UID: "4f99a167-975f-4373-8c9d-23a2ab6510fa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.124825 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24282d99-b526-4bd3-96b3-653642c42b9c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "24282d99-b526-4bd3-96b3-653642c42b9c" (UID: "24282d99-b526-4bd3-96b3-653642c42b9c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.135244 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f99a167-975f-4373-8c9d-23a2ab6510fa-kube-api-access-fgb9k" (OuterVolumeSpecName: "kube-api-access-fgb9k") pod "4f99a167-975f-4373-8c9d-23a2ab6510fa" (UID: "4f99a167-975f-4373-8c9d-23a2ab6510fa"). InnerVolumeSpecName "kube-api-access-fgb9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.135411 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff-kube-api-access-2sl5x" (OuterVolumeSpecName: "kube-api-access-2sl5x") pod "168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff" (UID: "168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff"). InnerVolumeSpecName "kube-api-access-2sl5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.136774 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06fe0f9a-dbc9-4078-92ce-07fe06eb4110-kube-api-access-jphwx" (OuterVolumeSpecName: "kube-api-access-jphwx") pod "06fe0f9a-dbc9-4078-92ce-07fe06eb4110" (UID: "06fe0f9a-dbc9-4078-92ce-07fe06eb4110"). InnerVolumeSpecName "kube-api-access-jphwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.139378 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d61a34c-59b3-490d-94a2-b9186426f179-kube-api-access-bthcr" (OuterVolumeSpecName: "kube-api-access-bthcr") pod "8d61a34c-59b3-490d-94a2-b9186426f179" (UID: "8d61a34c-59b3-490d-94a2-b9186426f179"). InnerVolumeSpecName "kube-api-access-bthcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.142752 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24282d99-b526-4bd3-96b3-653642c42b9c-kube-api-access-fp892" (OuterVolumeSpecName: "kube-api-access-fp892") pod "24282d99-b526-4bd3-96b3-653642c42b9c" (UID: "24282d99-b526-4bd3-96b3-653642c42b9c"). InnerVolumeSpecName "kube-api-access-fp892". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.149530 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/574bcef8-68d6-40d8-adbb-5a321a3b59ce-kube-api-access-z97fs" (OuterVolumeSpecName: "kube-api-access-z97fs") pod "574bcef8-68d6-40d8-adbb-5a321a3b59ce" (UID: "574bcef8-68d6-40d8-adbb-5a321a3b59ce"). InnerVolumeSpecName "kube-api-access-z97fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.225105 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jphwx\" (UniqueName: \"kubernetes.io/projected/06fe0f9a-dbc9-4078-92ce-07fe06eb4110-kube-api-access-jphwx\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.225143 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fp892\" (UniqueName: \"kubernetes.io/projected/24282d99-b526-4bd3-96b3-653642c42b9c-kube-api-access-fp892\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.225153 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.225162 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgb9k\" (UniqueName: \"kubernetes.io/projected/4f99a167-975f-4373-8c9d-23a2ab6510fa-kube-api-access-fgb9k\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.225172 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/24282d99-b526-4bd3-96b3-653642c42b9c-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.225181 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2sl5x\" (UniqueName: \"kubernetes.io/projected/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff-kube-api-access-2sl5x\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.225193 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/574bcef8-68d6-40d8-adbb-5a321a3b59ce-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.225202 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z97fs\" (UniqueName: \"kubernetes.io/projected/574bcef8-68d6-40d8-adbb-5a321a3b59ce-kube-api-access-z97fs\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.225210 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8d61a34c-59b3-490d-94a2-b9186426f179-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.225218 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06fe0f9a-dbc9-4078-92ce-07fe06eb4110-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.225227 4872 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4f99a167-975f-4373-8c9d-23a2ab6510fa-operator-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.225236 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bthcr\" (UniqueName: \"kubernetes.io/projected/8d61a34c-59b3-490d-94a2-b9186426f179-kube-api-access-bthcr\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:11 crc kubenswrapper[4872]: I0126 09:25:11.558225 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Jan 26 09:25:12 crc kubenswrapper[4872]: I0126 09:25:12.073631 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:25:12 crc kubenswrapper[4872]: I0126 09:25:12.073821 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:25:12 crc kubenswrapper[4872]: I0126 09:25:12.074574 4872 scope.go:117] "RemoveContainer" containerID="4257303280c6aa0818d7a3469f31ab0a87f3c6ef87fb5f5ae944311e72de51b6" Jan 26 09:25:12 crc kubenswrapper[4872]: E0126 09:25:12.075150 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-5b47d6fc7b-7wkw2_openstack(f721aeee-7daf-413d-885b-5c146881eb99)\"" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" podUID="f721aeee-7daf-413d-885b-5c146881eb99" Jan 26 09:25:12 crc kubenswrapper[4872]: I0126 09:25:12.113632 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerStarted","Data":"b7d9b838b0c2b8b8de7d52f9865ab4abf3484b60b57717140fdbe33bbfbdd6c5"} Jan 26 09:25:12 crc kubenswrapper[4872]: I0126 09:25:12.117516 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerStarted","Data":"4e5a56691abb3ad9e78de749bdd2887452395f6fa60f607182b0cb5f228434c9"} Jan 26 09:25:12 crc kubenswrapper[4872]: I0126 09:25:12.121347 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-4bea-account-create-update-vh8qs" Jan 26 09:25:12 crc kubenswrapper[4872]: I0126 09:25:12.122943 4872 scope.go:117] "RemoveContainer" containerID="4257303280c6aa0818d7a3469f31ab0a87f3c6ef87fb5f5ae944311e72de51b6" Jan 26 09:25:12 crc kubenswrapper[4872]: E0126 09:25:12.123777 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-5b47d6fc7b-7wkw2_openstack(f721aeee-7daf-413d-885b-5c146881eb99)\"" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" podUID="f721aeee-7daf-413d-885b-5c146881eb99" Jan 26 09:25:12 crc kubenswrapper[4872]: I0126 09:25:12.124262 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-bkbz5" Jan 26 09:25:12 crc kubenswrapper[4872]: I0126 09:25:12.124305 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cl2gt" Jan 26 09:25:12 crc kubenswrapper[4872]: I0126 09:25:12.124338 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5935-account-create-update-ftjg4" Jan 26 09:25:12 crc kubenswrapper[4872]: I0126 09:25:12.124368 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f91e-account-create-update-twns5" Jan 26 09:25:13 crc kubenswrapper[4872]: I0126 09:25:13.134749 4872 generic.go:334] "Generic (PLEG): container finished" podID="8865ee9d-9129-488e-84e2-cf1a11547f8a" containerID="b7d9b838b0c2b8b8de7d52f9865ab4abf3484b60b57717140fdbe33bbfbdd6c5" exitCode=0 Jan 26 09:25:13 crc kubenswrapper[4872]: I0126 09:25:13.134840 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerDied","Data":"b7d9b838b0c2b8b8de7d52f9865ab4abf3484b60b57717140fdbe33bbfbdd6c5"} Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.298808 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ws4hm"] Jan 26 09:25:14 crc kubenswrapper[4872]: E0126 09:25:14.299536 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24282d99-b526-4bd3-96b3-653642c42b9c" containerName="mariadb-database-create" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.299550 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="24282d99-b526-4bd3-96b3-653642c42b9c" containerName="mariadb-database-create" Jan 26 09:25:14 crc kubenswrapper[4872]: E0126 09:25:14.299562 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574bcef8-68d6-40d8-adbb-5a321a3b59ce" containerName="mariadb-account-create-update" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.299569 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="574bcef8-68d6-40d8-adbb-5a321a3b59ce" containerName="mariadb-account-create-update" Jan 26 09:25:14 crc kubenswrapper[4872]: E0126 09:25:14.299584 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06fe0f9a-dbc9-4078-92ce-07fe06eb4110" containerName="mariadb-account-create-update" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.299590 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="06fe0f9a-dbc9-4078-92ce-07fe06eb4110" containerName="mariadb-account-create-update" Jan 26 09:25:14 crc kubenswrapper[4872]: E0126 09:25:14.299612 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d61a34c-59b3-490d-94a2-b9186426f179" containerName="mariadb-account-create-update" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.299618 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d61a34c-59b3-490d-94a2-b9186426f179" containerName="mariadb-account-create-update" Jan 26 09:25:14 crc kubenswrapper[4872]: E0126 09:25:14.299639 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f99a167-975f-4373-8c9d-23a2ab6510fa" containerName="mariadb-database-create" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.299645 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f99a167-975f-4373-8c9d-23a2ab6510fa" containerName="mariadb-database-create" Jan 26 09:25:14 crc kubenswrapper[4872]: E0126 09:25:14.299660 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff" containerName="mariadb-database-create" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.299665 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff" containerName="mariadb-database-create" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.299852 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="06fe0f9a-dbc9-4078-92ce-07fe06eb4110" containerName="mariadb-account-create-update" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.299866 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f99a167-975f-4373-8c9d-23a2ab6510fa" containerName="mariadb-database-create" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.299886 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="24282d99-b526-4bd3-96b3-653642c42b9c" containerName="mariadb-database-create" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.299898 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="574bcef8-68d6-40d8-adbb-5a321a3b59ce" containerName="mariadb-account-create-update" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.299905 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d61a34c-59b3-490d-94a2-b9186426f179" containerName="mariadb-account-create-update" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.299915 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff" containerName="mariadb-database-create" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.300566 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.309058 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.309420 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.309598 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-grn5f" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.316174 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ws4hm"] Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.393679 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.393741 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.414519 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ws4hm\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.414607 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-config-data\") pod \"nova-cell0-conductor-db-sync-ws4hm\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.414651 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7flr\" (UniqueName: \"kubernetes.io/projected/f45318ef-cb2f-4816-b4c4-360b1f446f5c-kube-api-access-z7flr\") pod \"nova-cell0-conductor-db-sync-ws4hm\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.414694 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-scripts\") pod \"nova-cell0-conductor-db-sync-ws4hm\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.438936 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.442255 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.516498 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-config-data\") pod \"nova-cell0-conductor-db-sync-ws4hm\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.516584 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7flr\" (UniqueName: \"kubernetes.io/projected/f45318ef-cb2f-4816-b4c4-360b1f446f5c-kube-api-access-z7flr\") pod \"nova-cell0-conductor-db-sync-ws4hm\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.516665 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-scripts\") pod \"nova-cell0-conductor-db-sync-ws4hm\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.516778 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ws4hm\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.526513 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-scripts\") pod \"nova-cell0-conductor-db-sync-ws4hm\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.527262 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-ws4hm\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.527535 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-config-data\") pod \"nova-cell0-conductor-db-sync-ws4hm\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.534011 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.534061 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.548457 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7flr\" (UniqueName: \"kubernetes.io/projected/f45318ef-cb2f-4816-b4c4-360b1f446f5c-kube-api-access-z7flr\") pod \"nova-cell0-conductor-db-sync-ws4hm\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.613673 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.630111 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:14 crc kubenswrapper[4872]: I0126 09:25:14.633263 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:15 crc kubenswrapper[4872]: I0126 09:25:15.167342 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:15 crc kubenswrapper[4872]: I0126 09:25:15.167504 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 26 09:25:15 crc kubenswrapper[4872]: I0126 09:25:15.167673 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Jan 26 09:25:15 crc kubenswrapper[4872]: I0126 09:25:15.167887 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:16 crc kubenswrapper[4872]: I0126 09:25:16.857433 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ws4hm"] Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.214223 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ws4hm" event={"ID":"f45318ef-cb2f-4816-b4c4-360b1f446f5c","Type":"ContainerStarted","Data":"5f6357ba48698a25d2e1b902896073953d60040d61bc808210200e225f983fce"} Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.222598 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"52a92868-e230-4403-9297-51c2b457343b","Type":"ContainerStarted","Data":"38031c56859755cfa2f31ab75e1c5830d060932bdc5ede8db8db8c42301b9592"} Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.224334 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.231403 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"e135958c-61c2-4199-94b3-2a8e6623310c","Type":"ContainerStarted","Data":"7d42cdc2ea0cb003d41f0df8602b37aab9808928402483a19fbde86b805272f0"} Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.233600 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.236651 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.239870 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerStarted","Data":"09b82b9161ffa8b472244763f5817478abbdff26875e400a782feb5c4671c15f"} Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.240326 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.240347 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.240537 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.265072 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.716319654 podStartE2EDuration="39.265041807s" podCreationTimestamp="2026-01-26 09:24:38 +0000 UTC" firstStartedPulling="2026-01-26 09:24:39.889951347 +0000 UTC m=+1013.198791148" lastFinishedPulling="2026-01-26 09:25:16.4386735 +0000 UTC m=+1049.747513301" observedRunningTime="2026-01-26 09:25:17.251506529 +0000 UTC m=+1050.560346330" watchObservedRunningTime="2026-01-26 09:25:17.265041807 +0000 UTC m=+1050.573881608" Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.423898 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:17 crc kubenswrapper[4872]: I0126 09:25:17.430080 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Jan 26 09:25:18 crc kubenswrapper[4872]: I0126 09:25:18.280380 4872 generic.go:334] "Generic (PLEG): container finished" podID="8865ee9d-9129-488e-84e2-cf1a11547f8a" containerID="09b82b9161ffa8b472244763f5817478abbdff26875e400a782feb5c4671c15f" exitCode=0 Jan 26 09:25:18 crc kubenswrapper[4872]: I0126 09:25:18.280450 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerDied","Data":"09b82b9161ffa8b472244763f5817478abbdff26875e400a782feb5c4671c15f"} Jan 26 09:25:19 crc kubenswrapper[4872]: I0126 09:25:19.296219 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerStarted","Data":"5e174e12751de1f8ad21e27c5b24f46e3a4a25ca7c6eb74d35181e392b9e33c5"} Jan 26 09:25:20 crc kubenswrapper[4872]: I0126 09:25:20.327153 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerStarted","Data":"3c756481c2c8407b86488d715d3de72dcd671affb26735e51b02cb3533a15141"} Jan 26 09:25:21 crc kubenswrapper[4872]: I0126 09:25:21.357815 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerStarted","Data":"bf4d9a9144b8ece1fdd357e5cfb10ece9f101ee886283858cd4a29b5099e979e"} Jan 26 09:25:22 crc kubenswrapper[4872]: I0126 09:25:22.374350 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerStarted","Data":"a386f7970f7079e01a0415634f90d0acf172e2f4c70a13017b1ca725f05c6bd9"} Jan 26 09:25:22 crc kubenswrapper[4872]: I0126 09:25:22.375240 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Jan 26 09:25:22 crc kubenswrapper[4872]: I0126 09:25:22.421770 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-inspector-0" podStartSLOduration=10.098881001 podStartE2EDuration="13.421739346s" podCreationTimestamp="2026-01-26 09:25:09 +0000 UTC" firstStartedPulling="2026-01-26 09:25:13.139488403 +0000 UTC m=+1046.448328204" lastFinishedPulling="2026-01-26 09:25:16.462346738 +0000 UTC m=+1049.771186549" observedRunningTime="2026-01-26 09:25:22.407219993 +0000 UTC m=+1055.716059804" watchObservedRunningTime="2026-01-26 09:25:22.421739346 +0000 UTC m=+1055.730579157" Jan 26 09:25:23 crc kubenswrapper[4872]: I0126 09:25:23.389322 4872 generic.go:334] "Generic (PLEG): container finished" podID="8865ee9d-9129-488e-84e2-cf1a11547f8a" containerID="3c756481c2c8407b86488d715d3de72dcd671affb26735e51b02cb3533a15141" exitCode=0 Jan 26 09:25:23 crc kubenswrapper[4872]: I0126 09:25:23.389822 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerDied","Data":"3c756481c2c8407b86488d715d3de72dcd671affb26735e51b02cb3533a15141"} Jan 26 09:25:23 crc kubenswrapper[4872]: I0126 09:25:23.390435 4872 scope.go:117] "RemoveContainer" containerID="3c756481c2c8407b86488d715d3de72dcd671affb26735e51b02cb3533a15141" Jan 26 09:25:24 crc kubenswrapper[4872]: I0126 09:25:24.588856 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Jan 26 09:25:24 crc kubenswrapper[4872]: I0126 09:25:24.588932 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Jan 26 09:25:24 crc kubenswrapper[4872]: I0126 09:25:24.588966 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-inspector-0" Jan 26 09:25:27 crc kubenswrapper[4872]: I0126 09:25:27.199952 4872 scope.go:117] "RemoveContainer" containerID="4257303280c6aa0818d7a3469f31ab0a87f3c6ef87fb5f5ae944311e72de51b6" Jan 26 09:25:27 crc kubenswrapper[4872]: E0126 09:25:27.201760 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-5b47d6fc7b-7wkw2_openstack(f721aeee-7daf-413d-885b-5c146881eb99)\"" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" podUID="f721aeee-7daf-413d-885b-5c146881eb99" Jan 26 09:25:29 crc kubenswrapper[4872]: I0126 09:25:29.476312 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerStarted","Data":"8785fa027c424ae1f130f2bad65d60674cbf0f975ff79a3a0048afbab6984c21"} Jan 26 09:25:29 crc kubenswrapper[4872]: I0126 09:25:29.483086 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ws4hm" event={"ID":"f45318ef-cb2f-4816-b4c4-360b1f446f5c","Type":"ContainerStarted","Data":"1bb465fcc2f76c703eac83dc4771aa923acb2c20161bb28419cb57ad318181ab"} Jan 26 09:25:29 crc kubenswrapper[4872]: I0126 09:25:29.484233 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Jan 26 09:25:29 crc kubenswrapper[4872]: I0126 09:25:29.544976 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-ws4hm" podStartSLOduration=3.957457185 podStartE2EDuration="15.544949145s" podCreationTimestamp="2026-01-26 09:25:14 +0000 UTC" firstStartedPulling="2026-01-26 09:25:16.875064754 +0000 UTC m=+1050.183904555" lastFinishedPulling="2026-01-26 09:25:28.462556714 +0000 UTC m=+1061.771396515" observedRunningTime="2026-01-26 09:25:29.535665807 +0000 UTC m=+1062.844505628" watchObservedRunningTime="2026-01-26 09:25:29.544949145 +0000 UTC m=+1062.853788946" Jan 26 09:25:29 crc kubenswrapper[4872]: I0126 09:25:29.588820 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Jan 26 09:25:29 crc kubenswrapper[4872]: I0126 09:25:29.588908 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-inspector-0" Jan 26 09:25:29 crc kubenswrapper[4872]: I0126 09:25:29.588945 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-inspector-0" Jan 26 09:25:29 crc kubenswrapper[4872]: I0126 09:25:29.597767 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/ironic-inspector-0" podUID="8865ee9d-9129-488e-84e2-cf1a11547f8a" containerName="ironic-inspector-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Jan 26 09:25:29 crc kubenswrapper[4872]: I0126 09:25:29.598284 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/ironic-inspector-0" podUID="8865ee9d-9129-488e-84e2-cf1a11547f8a" containerName="ironic-inspector" probeResult="failure" output="HTTP probe failed with statuscode: 503" Jan 26 09:25:32 crc kubenswrapper[4872]: I0126 09:25:32.522501 4872 generic.go:334] "Generic (PLEG): container finished" podID="8865ee9d-9129-488e-84e2-cf1a11547f8a" containerID="8785fa027c424ae1f130f2bad65d60674cbf0f975ff79a3a0048afbab6984c21" exitCode=0 Jan 26 09:25:32 crc kubenswrapper[4872]: I0126 09:25:32.522582 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerDied","Data":"8785fa027c424ae1f130f2bad65d60674cbf0f975ff79a3a0048afbab6984c21"} Jan 26 09:25:32 crc kubenswrapper[4872]: I0126 09:25:32.523136 4872 scope.go:117] "RemoveContainer" containerID="3c756481c2c8407b86488d715d3de72dcd671affb26735e51b02cb3533a15141" Jan 26 09:25:32 crc kubenswrapper[4872]: I0126 09:25:32.524200 4872 scope.go:117] "RemoveContainer" containerID="8785fa027c424ae1f130f2bad65d60674cbf0f975ff79a3a0048afbab6984c21" Jan 26 09:25:32 crc kubenswrapper[4872]: E0126 09:25:32.524632 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-inspector\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-inspector pod=ironic-inspector-0_openstack(8865ee9d-9129-488e-84e2-cf1a11547f8a)\"" pod="openstack/ironic-inspector-0" podUID="8865ee9d-9129-488e-84e2-cf1a11547f8a" Jan 26 09:25:34 crc kubenswrapper[4872]: I0126 09:25:34.588529 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-inspector-0" Jan 26 09:25:34 crc kubenswrapper[4872]: I0126 09:25:34.590416 4872 scope.go:117] "RemoveContainer" containerID="8785fa027c424ae1f130f2bad65d60674cbf0f975ff79a3a0048afbab6984c21" Jan 26 09:25:34 crc kubenswrapper[4872]: E0126 09:25:34.590777 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-inspector\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-inspector pod=ironic-inspector-0_openstack(8865ee9d-9129-488e-84e2-cf1a11547f8a)\"" pod="openstack/ironic-inspector-0" podUID="8865ee9d-9129-488e-84e2-cf1a11547f8a" Jan 26 09:25:36 crc kubenswrapper[4872]: I0126 09:25:36.575825 4872 generic.go:334] "Generic (PLEG): container finished" podID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerID="3c6e9a0f17a77512158f836fd07c3ed0b7809f2a23434fd3161ed21bd50540cb" exitCode=137 Jan 26 09:25:36 crc kubenswrapper[4872]: I0126 09:25:36.575902 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94a8957b-b742-4dfd-a463-e265cf4db32e","Type":"ContainerDied","Data":"3c6e9a0f17a77512158f836fd07c3ed0b7809f2a23434fd3161ed21bd50540cb"} Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.340002 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.463273 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94a8957b-b742-4dfd-a463-e265cf4db32e-run-httpd\") pod \"94a8957b-b742-4dfd-a463-e265cf4db32e\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.463380 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-config-data\") pod \"94a8957b-b742-4dfd-a463-e265cf4db32e\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.463447 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-scripts\") pod \"94a8957b-b742-4dfd-a463-e265cf4db32e\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.463601 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-combined-ca-bundle\") pod \"94a8957b-b742-4dfd-a463-e265cf4db32e\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.463633 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l5p2\" (UniqueName: \"kubernetes.io/projected/94a8957b-b742-4dfd-a463-e265cf4db32e-kube-api-access-6l5p2\") pod \"94a8957b-b742-4dfd-a463-e265cf4db32e\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.463698 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-sg-core-conf-yaml\") pod \"94a8957b-b742-4dfd-a463-e265cf4db32e\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.463782 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94a8957b-b742-4dfd-a463-e265cf4db32e-log-httpd\") pod \"94a8957b-b742-4dfd-a463-e265cf4db32e\" (UID: \"94a8957b-b742-4dfd-a463-e265cf4db32e\") " Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.464744 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94a8957b-b742-4dfd-a463-e265cf4db32e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "94a8957b-b742-4dfd-a463-e265cf4db32e" (UID: "94a8957b-b742-4dfd-a463-e265cf4db32e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.464934 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94a8957b-b742-4dfd-a463-e265cf4db32e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "94a8957b-b742-4dfd-a463-e265cf4db32e" (UID: "94a8957b-b742-4dfd-a463-e265cf4db32e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.476485 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-scripts" (OuterVolumeSpecName: "scripts") pod "94a8957b-b742-4dfd-a463-e265cf4db32e" (UID: "94a8957b-b742-4dfd-a463-e265cf4db32e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.488432 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94a8957b-b742-4dfd-a463-e265cf4db32e-kube-api-access-6l5p2" (OuterVolumeSpecName: "kube-api-access-6l5p2") pod "94a8957b-b742-4dfd-a463-e265cf4db32e" (UID: "94a8957b-b742-4dfd-a463-e265cf4db32e"). InnerVolumeSpecName "kube-api-access-6l5p2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.501065 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "94a8957b-b742-4dfd-a463-e265cf4db32e" (UID: "94a8957b-b742-4dfd-a463-e265cf4db32e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.565905 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94a8957b-b742-4dfd-a463-e265cf4db32e-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.565946 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/94a8957b-b742-4dfd-a463-e265cf4db32e-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.565959 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.565972 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l5p2\" (UniqueName: \"kubernetes.io/projected/94a8957b-b742-4dfd-a463-e265cf4db32e-kube-api-access-6l5p2\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.565988 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.565975 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94a8957b-b742-4dfd-a463-e265cf4db32e" (UID: "94a8957b-b742-4dfd-a463-e265cf4db32e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.585931 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-config-data" (OuterVolumeSpecName: "config-data") pod "94a8957b-b742-4dfd-a463-e265cf4db32e" (UID: "94a8957b-b742-4dfd-a463-e265cf4db32e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.591456 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"94a8957b-b742-4dfd-a463-e265cf4db32e","Type":"ContainerDied","Data":"8e256f948be1dfdc015f44d318ec72c7e2c3b157c8c0e5b7bbad072f0c919a9d"} Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.591551 4872 scope.go:117] "RemoveContainer" containerID="3c6e9a0f17a77512158f836fd07c3ed0b7809f2a23434fd3161ed21bd50540cb" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.591867 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.657355 4872 scope.go:117] "RemoveContainer" containerID="2950135b35b063ebe45e02bfcda7adcaf3a0484a7845bb34ea8eeb9a995655e3" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.665913 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.667918 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.667956 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94a8957b-b742-4dfd-a463-e265cf4db32e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.676046 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.677179 4872 scope.go:117] "RemoveContainer" containerID="f850eae1ba4ede05658e235af04af5e9ddc091ac135bdddf5732952720385fd6" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.701624 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:25:37 crc kubenswrapper[4872]: E0126 09:25:37.702150 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="proxy-httpd" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.702177 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="proxy-httpd" Jan 26 09:25:37 crc kubenswrapper[4872]: E0126 09:25:37.702201 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="ceilometer-central-agent" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.702209 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="ceilometer-central-agent" Jan 26 09:25:37 crc kubenswrapper[4872]: E0126 09:25:37.702220 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="ceilometer-notification-agent" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.702227 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="ceilometer-notification-agent" Jan 26 09:25:37 crc kubenswrapper[4872]: E0126 09:25:37.702260 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="sg-core" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.702267 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="sg-core" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.702460 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="sg-core" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.702492 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="ceilometer-notification-agent" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.702516 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="proxy-httpd" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.702529 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" containerName="ceilometer-central-agent" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.704699 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.707777 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.707855 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.715608 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.729158 4872 scope.go:117] "RemoveContainer" containerID="2272551f1062c8f5581604bfa484ed493b479d73101aa07319ce08c2141898ee" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.876652 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-scripts\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.877100 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-config-data\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.877256 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.877444 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6lkb\" (UniqueName: \"kubernetes.io/projected/ca66fbcd-6af9-42eb-82a1-1e42268194c8-kube-api-access-q6lkb\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.877583 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca66fbcd-6af9-42eb-82a1-1e42268194c8-run-httpd\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.877698 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.877875 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca66fbcd-6af9-42eb-82a1-1e42268194c8-log-httpd\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.980443 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-config-data\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.980532 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.980623 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6lkb\" (UniqueName: \"kubernetes.io/projected/ca66fbcd-6af9-42eb-82a1-1e42268194c8-kube-api-access-q6lkb\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.980694 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca66fbcd-6af9-42eb-82a1-1e42268194c8-run-httpd\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.980730 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.980832 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca66fbcd-6af9-42eb-82a1-1e42268194c8-log-httpd\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.980939 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-scripts\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.981400 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca66fbcd-6af9-42eb-82a1-1e42268194c8-run-httpd\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:37 crc kubenswrapper[4872]: I0126 09:25:37.981570 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca66fbcd-6af9-42eb-82a1-1e42268194c8-log-httpd\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:38 crc kubenswrapper[4872]: I0126 09:25:38.000142 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:38 crc kubenswrapper[4872]: I0126 09:25:38.000465 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:38 crc kubenswrapper[4872]: I0126 09:25:38.000903 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-config-data\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:38 crc kubenswrapper[4872]: I0126 09:25:38.004320 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-scripts\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:38 crc kubenswrapper[4872]: I0126 09:25:38.005550 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6lkb\" (UniqueName: \"kubernetes.io/projected/ca66fbcd-6af9-42eb-82a1-1e42268194c8-kube-api-access-q6lkb\") pod \"ceilometer-0\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " pod="openstack/ceilometer-0" Jan 26 09:25:38 crc kubenswrapper[4872]: I0126 09:25:38.031256 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:25:38 crc kubenswrapper[4872]: I0126 09:25:38.536704 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:25:38 crc kubenswrapper[4872]: W0126 09:25:38.543893 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca66fbcd_6af9_42eb_82a1_1e42268194c8.slice/crio-d0f008a59a802e22bf89c95d74feda5101b0340e35043af075297898dae767e4 WatchSource:0}: Error finding container d0f008a59a802e22bf89c95d74feda5101b0340e35043af075297898dae767e4: Status 404 returned error can't find the container with id d0f008a59a802e22bf89c95d74feda5101b0340e35043af075297898dae767e4 Jan 26 09:25:38 crc kubenswrapper[4872]: I0126 09:25:38.604662 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca66fbcd-6af9-42eb-82a1-1e42268194c8","Type":"ContainerStarted","Data":"d0f008a59a802e22bf89c95d74feda5101b0340e35043af075297898dae767e4"} Jan 26 09:25:39 crc kubenswrapper[4872]: I0126 09:25:39.197139 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94a8957b-b742-4dfd-a463-e265cf4db32e" path="/var/lib/kubelet/pods/94a8957b-b742-4dfd-a463-e265cf4db32e/volumes" Jan 26 09:25:39 crc kubenswrapper[4872]: I0126 09:25:39.588790 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-inspector-0" Jan 26 09:25:39 crc kubenswrapper[4872]: I0126 09:25:39.590521 4872 scope.go:117] "RemoveContainer" containerID="8785fa027c424ae1f130f2bad65d60674cbf0f975ff79a3a0048afbab6984c21" Jan 26 09:25:39 crc kubenswrapper[4872]: E0126 09:25:39.590839 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-inspector\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-inspector pod=ironic-inspector-0_openstack(8865ee9d-9129-488e-84e2-cf1a11547f8a)\"" pod="openstack/ironic-inspector-0" podUID="8865ee9d-9129-488e-84e2-cf1a11547f8a" Jan 26 09:25:39 crc kubenswrapper[4872]: I0126 09:25:39.597755 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/ironic-inspector-0" podUID="8865ee9d-9129-488e-84e2-cf1a11547f8a" containerName="ironic-inspector-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Jan 26 09:25:39 crc kubenswrapper[4872]: I0126 09:25:39.620529 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca66fbcd-6af9-42eb-82a1-1e42268194c8","Type":"ContainerStarted","Data":"84dab645179878c106b0a586bd291b127e672af3857d1d73b91f24d160a3d99f"} Jan 26 09:25:40 crc kubenswrapper[4872]: I0126 09:25:40.633749 4872 generic.go:334] "Generic (PLEG): container finished" podID="f45318ef-cb2f-4816-b4c4-360b1f446f5c" containerID="1bb465fcc2f76c703eac83dc4771aa923acb2c20161bb28419cb57ad318181ab" exitCode=0 Jan 26 09:25:40 crc kubenswrapper[4872]: I0126 09:25:40.633855 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ws4hm" event={"ID":"f45318ef-cb2f-4816-b4c4-360b1f446f5c","Type":"ContainerDied","Data":"1bb465fcc2f76c703eac83dc4771aa923acb2c20161bb28419cb57ad318181ab"} Jan 26 09:25:40 crc kubenswrapper[4872]: I0126 09:25:40.637305 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca66fbcd-6af9-42eb-82a1-1e42268194c8","Type":"ContainerStarted","Data":"7fce3bb5e9d605d39dabf5d200b81991990c1d326759fc6fbc810469768259fa"} Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.094047 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.184753 4872 scope.go:117] "RemoveContainer" containerID="4257303280c6aa0818d7a3469f31ab0a87f3c6ef87fb5f5ae944311e72de51b6" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.295449 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-combined-ca-bundle\") pod \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.295651 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-config-data\") pod \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.295786 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-scripts\") pod \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.295924 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7flr\" (UniqueName: \"kubernetes.io/projected/f45318ef-cb2f-4816-b4c4-360b1f446f5c-kube-api-access-z7flr\") pod \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\" (UID: \"f45318ef-cb2f-4816-b4c4-360b1f446f5c\") " Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.301944 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-scripts" (OuterVolumeSpecName: "scripts") pod "f45318ef-cb2f-4816-b4c4-360b1f446f5c" (UID: "f45318ef-cb2f-4816-b4c4-360b1f446f5c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.302514 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f45318ef-cb2f-4816-b4c4-360b1f446f5c-kube-api-access-z7flr" (OuterVolumeSpecName: "kube-api-access-z7flr") pod "f45318ef-cb2f-4816-b4c4-360b1f446f5c" (UID: "f45318ef-cb2f-4816-b4c4-360b1f446f5c"). InnerVolumeSpecName "kube-api-access-z7flr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.333468 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f45318ef-cb2f-4816-b4c4-360b1f446f5c" (UID: "f45318ef-cb2f-4816-b4c4-360b1f446f5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.341533 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-config-data" (OuterVolumeSpecName: "config-data") pod "f45318ef-cb2f-4816-b4c4-360b1f446f5c" (UID: "f45318ef-cb2f-4816-b4c4-360b1f446f5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.399167 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.399206 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7flr\" (UniqueName: \"kubernetes.io/projected/f45318ef-cb2f-4816-b4c4-360b1f446f5c-kube-api-access-z7flr\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.399222 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.399231 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f45318ef-cb2f-4816-b4c4-360b1f446f5c-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.671330 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca66fbcd-6af9-42eb-82a1-1e42268194c8","Type":"ContainerStarted","Data":"e5cdda80ae51aa0623048835faa1c7f7a6074290a97a27e5e32420143e313b33"} Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.673498 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-ws4hm" event={"ID":"f45318ef-cb2f-4816-b4c4-360b1f446f5c","Type":"ContainerDied","Data":"5f6357ba48698a25d2e1b902896073953d60040d61bc808210200e225f983fce"} Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.673532 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f6357ba48698a25d2e1b902896073953d60040d61bc808210200e225f983fce" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.673597 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-ws4hm" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.840503 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 26 09:25:42 crc kubenswrapper[4872]: E0126 09:25:42.841082 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f45318ef-cb2f-4816-b4c4-360b1f446f5c" containerName="nova-cell0-conductor-db-sync" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.841112 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f45318ef-cb2f-4816-b4c4-360b1f446f5c" containerName="nova-cell0-conductor-db-sync" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.841388 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f45318ef-cb2f-4816-b4c4-360b1f446f5c" containerName="nova-cell0-conductor-db-sync" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.842265 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.847989 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.848142 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-grn5f" Jan 26 09:25:42 crc kubenswrapper[4872]: I0126 09:25:42.886575 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.012018 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34fd0f4-3e8c-42a8-a242-5c0bea758f42-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d34fd0f4-3e8c-42a8-a242-5c0bea758f42\") " pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.012086 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34fd0f4-3e8c-42a8-a242-5c0bea758f42-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d34fd0f4-3e8c-42a8-a242-5c0bea758f42\") " pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.012189 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw8wp\" (UniqueName: \"kubernetes.io/projected/d34fd0f4-3e8c-42a8-a242-5c0bea758f42-kube-api-access-tw8wp\") pod \"nova-cell0-conductor-0\" (UID: \"d34fd0f4-3e8c-42a8-a242-5c0bea758f42\") " pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.114585 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw8wp\" (UniqueName: \"kubernetes.io/projected/d34fd0f4-3e8c-42a8-a242-5c0bea758f42-kube-api-access-tw8wp\") pod \"nova-cell0-conductor-0\" (UID: \"d34fd0f4-3e8c-42a8-a242-5c0bea758f42\") " pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.115024 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34fd0f4-3e8c-42a8-a242-5c0bea758f42-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d34fd0f4-3e8c-42a8-a242-5c0bea758f42\") " pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.115641 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34fd0f4-3e8c-42a8-a242-5c0bea758f42-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d34fd0f4-3e8c-42a8-a242-5c0bea758f42\") " pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.118861 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d34fd0f4-3e8c-42a8-a242-5c0bea758f42-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"d34fd0f4-3e8c-42a8-a242-5c0bea758f42\") " pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.119358 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d34fd0f4-3e8c-42a8-a242-5c0bea758f42-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"d34fd0f4-3e8c-42a8-a242-5c0bea758f42\") " pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.147363 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw8wp\" (UniqueName: \"kubernetes.io/projected/d34fd0f4-3e8c-42a8-a242-5c0bea758f42-kube-api-access-tw8wp\") pod \"nova-cell0-conductor-0\" (UID: \"d34fd0f4-3e8c-42a8-a242-5c0bea758f42\") " pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.176769 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.683985 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" event={"ID":"f721aeee-7daf-413d-885b-5c146881eb99","Type":"ContainerStarted","Data":"c46f273a2c765b8c93d1ba6ff7456a51f5636420af41ec4031de973f45e9165d"} Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.686332 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:25:43 crc kubenswrapper[4872]: I0126 09:25:43.713258 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Jan 26 09:25:44 crc kubenswrapper[4872]: I0126 09:25:44.696087 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca66fbcd-6af9-42eb-82a1-1e42268194c8","Type":"ContainerStarted","Data":"c52ad9f20ed656c9b4033d19ae7c49c942d15e6384be3ecd730b064571bb6407"} Jan 26 09:25:44 crc kubenswrapper[4872]: I0126 09:25:44.696703 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 09:25:44 crc kubenswrapper[4872]: I0126 09:25:44.699784 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d34fd0f4-3e8c-42a8-a242-5c0bea758f42","Type":"ContainerStarted","Data":"54a278761b668e0faa384421f2886ab08f0f148bb024e05201fba62bf1e30f3c"} Jan 26 09:25:44 crc kubenswrapper[4872]: I0126 09:25:44.699846 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"d34fd0f4-3e8c-42a8-a242-5c0bea758f42","Type":"ContainerStarted","Data":"c2052aa2e017c9e65e1312ece7bf7e174f6a335724a584b64aac11dfc0293869"} Jan 26 09:25:44 crc kubenswrapper[4872]: I0126 09:25:44.699859 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:44 crc kubenswrapper[4872]: I0126 09:25:44.723152 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.798408745 podStartE2EDuration="7.723129328s" podCreationTimestamp="2026-01-26 09:25:37 +0000 UTC" firstStartedPulling="2026-01-26 09:25:38.548124804 +0000 UTC m=+1071.856964605" lastFinishedPulling="2026-01-26 09:25:43.472845387 +0000 UTC m=+1076.781685188" observedRunningTime="2026-01-26 09:25:44.718935051 +0000 UTC m=+1078.027774852" watchObservedRunningTime="2026-01-26 09:25:44.723129328 +0000 UTC m=+1078.031969129" Jan 26 09:25:44 crc kubenswrapper[4872]: I0126 09:25:44.740061 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.740036553 podStartE2EDuration="2.740036553s" podCreationTimestamp="2026-01-26 09:25:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:25:44.737841616 +0000 UTC m=+1078.046681417" watchObservedRunningTime="2026-01-26 09:25:44.740036553 +0000 UTC m=+1078.048876354" Jan 26 09:25:46 crc kubenswrapper[4872]: I0126 09:25:46.136899 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:25:46 crc kubenswrapper[4872]: I0126 09:25:46.720455 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="ceilometer-central-agent" containerID="cri-o://84dab645179878c106b0a586bd291b127e672af3857d1d73b91f24d160a3d99f" gracePeriod=30 Jan 26 09:25:46 crc kubenswrapper[4872]: I0126 09:25:46.721149 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="ceilometer-notification-agent" containerID="cri-o://7fce3bb5e9d605d39dabf5d200b81991990c1d326759fc6fbc810469768259fa" gracePeriod=30 Jan 26 09:25:46 crc kubenswrapper[4872]: I0126 09:25:46.721048 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="proxy-httpd" containerID="cri-o://c52ad9f20ed656c9b4033d19ae7c49c942d15e6384be3ecd730b064571bb6407" gracePeriod=30 Jan 26 09:25:46 crc kubenswrapper[4872]: I0126 09:25:46.721147 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="sg-core" containerID="cri-o://e5cdda80ae51aa0623048835faa1c7f7a6074290a97a27e5e32420143e313b33" gracePeriod=30 Jan 26 09:25:47 crc kubenswrapper[4872]: I0126 09:25:47.120763 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-neutron-agent-5b47d6fc7b-7wkw2" Jan 26 09:25:47 crc kubenswrapper[4872]: I0126 09:25:47.731373 4872 generic.go:334] "Generic (PLEG): container finished" podID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerID="c52ad9f20ed656c9b4033d19ae7c49c942d15e6384be3ecd730b064571bb6407" exitCode=0 Jan 26 09:25:47 crc kubenswrapper[4872]: I0126 09:25:47.731898 4872 generic.go:334] "Generic (PLEG): container finished" podID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerID="e5cdda80ae51aa0623048835faa1c7f7a6074290a97a27e5e32420143e313b33" exitCode=2 Jan 26 09:25:47 crc kubenswrapper[4872]: I0126 09:25:47.731911 4872 generic.go:334] "Generic (PLEG): container finished" podID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerID="7fce3bb5e9d605d39dabf5d200b81991990c1d326759fc6fbc810469768259fa" exitCode=0 Jan 26 09:25:47 crc kubenswrapper[4872]: I0126 09:25:47.731448 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca66fbcd-6af9-42eb-82a1-1e42268194c8","Type":"ContainerDied","Data":"c52ad9f20ed656c9b4033d19ae7c49c942d15e6384be3ecd730b064571bb6407"} Jan 26 09:25:47 crc kubenswrapper[4872]: I0126 09:25:47.731959 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca66fbcd-6af9-42eb-82a1-1e42268194c8","Type":"ContainerDied","Data":"e5cdda80ae51aa0623048835faa1c7f7a6074290a97a27e5e32420143e313b33"} Jan 26 09:25:47 crc kubenswrapper[4872]: I0126 09:25:47.731980 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca66fbcd-6af9-42eb-82a1-1e42268194c8","Type":"ContainerDied","Data":"7fce3bb5e9d605d39dabf5d200b81991990c1d326759fc6fbc810469768259fa"} Jan 26 09:25:49 crc kubenswrapper[4872]: I0126 09:25:49.597627 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/ironic-inspector-0" podUID="8865ee9d-9129-488e-84e2-cf1a11547f8a" containerName="ironic-inspector-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Jan 26 09:25:51 crc kubenswrapper[4872]: I0126 09:25:51.185085 4872 scope.go:117] "RemoveContainer" containerID="8785fa027c424ae1f130f2bad65d60674cbf0f975ff79a3a0048afbab6984c21" Jan 26 09:25:51 crc kubenswrapper[4872]: I0126 09:25:51.782586 4872 generic.go:334] "Generic (PLEG): container finished" podID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerID="84dab645179878c106b0a586bd291b127e672af3857d1d73b91f24d160a3d99f" exitCode=0 Jan 26 09:25:51 crc kubenswrapper[4872]: I0126 09:25:51.782691 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca66fbcd-6af9-42eb-82a1-1e42268194c8","Type":"ContainerDied","Data":"84dab645179878c106b0a586bd291b127e672af3857d1d73b91f24d160a3d99f"} Jan 26 09:25:51 crc kubenswrapper[4872]: I0126 09:25:51.791622 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"8865ee9d-9129-488e-84e2-cf1a11547f8a","Type":"ContainerStarted","Data":"41fdc2eaec0a2fecd11bc08e2de16b8ba9368e49d7072b74db09d459e1e111ef"} Jan 26 09:25:51 crc kubenswrapper[4872]: I0126 09:25:51.931142 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.121536 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-sg-core-conf-yaml\") pod \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.121652 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-config-data\") pod \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.121849 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6lkb\" (UniqueName: \"kubernetes.io/projected/ca66fbcd-6af9-42eb-82a1-1e42268194c8-kube-api-access-q6lkb\") pod \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.121877 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca66fbcd-6af9-42eb-82a1-1e42268194c8-log-httpd\") pod \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.121907 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-scripts\") pod \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.121978 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-combined-ca-bundle\") pod \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.122029 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca66fbcd-6af9-42eb-82a1-1e42268194c8-run-httpd\") pod \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\" (UID: \"ca66fbcd-6af9-42eb-82a1-1e42268194c8\") " Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.122952 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca66fbcd-6af9-42eb-82a1-1e42268194c8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ca66fbcd-6af9-42eb-82a1-1e42268194c8" (UID: "ca66fbcd-6af9-42eb-82a1-1e42268194c8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.123145 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca66fbcd-6af9-42eb-82a1-1e42268194c8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ca66fbcd-6af9-42eb-82a1-1e42268194c8" (UID: "ca66fbcd-6af9-42eb-82a1-1e42268194c8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.124216 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca66fbcd-6af9-42eb-82a1-1e42268194c8-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.124230 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ca66fbcd-6af9-42eb-82a1-1e42268194c8-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.130215 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca66fbcd-6af9-42eb-82a1-1e42268194c8-kube-api-access-q6lkb" (OuterVolumeSpecName: "kube-api-access-q6lkb") pod "ca66fbcd-6af9-42eb-82a1-1e42268194c8" (UID: "ca66fbcd-6af9-42eb-82a1-1e42268194c8"). InnerVolumeSpecName "kube-api-access-q6lkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.130936 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-scripts" (OuterVolumeSpecName: "scripts") pod "ca66fbcd-6af9-42eb-82a1-1e42268194c8" (UID: "ca66fbcd-6af9-42eb-82a1-1e42268194c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.152262 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ca66fbcd-6af9-42eb-82a1-1e42268194c8" (UID: "ca66fbcd-6af9-42eb-82a1-1e42268194c8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.220214 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca66fbcd-6af9-42eb-82a1-1e42268194c8" (UID: "ca66fbcd-6af9-42eb-82a1-1e42268194c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.226813 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6lkb\" (UniqueName: \"kubernetes.io/projected/ca66fbcd-6af9-42eb-82a1-1e42268194c8-kube-api-access-q6lkb\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.226847 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.226866 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.226878 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.231937 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-config-data" (OuterVolumeSpecName: "config-data") pod "ca66fbcd-6af9-42eb-82a1-1e42268194c8" (UID: "ca66fbcd-6af9-42eb-82a1-1e42268194c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.329600 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca66fbcd-6af9-42eb-82a1-1e42268194c8-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.817549 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ca66fbcd-6af9-42eb-82a1-1e42268194c8","Type":"ContainerDied","Data":"d0f008a59a802e22bf89c95d74feda5101b0340e35043af075297898dae767e4"} Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.817659 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.818149 4872 scope.go:117] "RemoveContainer" containerID="c52ad9f20ed656c9b4033d19ae7c49c942d15e6384be3ecd730b064571bb6407" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.880591 4872 scope.go:117] "RemoveContainer" containerID="e5cdda80ae51aa0623048835faa1c7f7a6074290a97a27e5e32420143e313b33" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.905712 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.919922 4872 scope.go:117] "RemoveContainer" containerID="7fce3bb5e9d605d39dabf5d200b81991990c1d326759fc6fbc810469768259fa" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.945289 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.949573 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:25:52 crc kubenswrapper[4872]: E0126 09:25:52.950387 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="ceilometer-central-agent" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.950425 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="ceilometer-central-agent" Jan 26 09:25:52 crc kubenswrapper[4872]: E0126 09:25:52.950462 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="proxy-httpd" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.950477 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="proxy-httpd" Jan 26 09:25:52 crc kubenswrapper[4872]: E0126 09:25:52.950499 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="ceilometer-notification-agent" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.950513 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="ceilometer-notification-agent" Jan 26 09:25:52 crc kubenswrapper[4872]: E0126 09:25:52.950552 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="sg-core" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.950566 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="sg-core" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.950955 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="sg-core" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.951000 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="ceilometer-notification-agent" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.951024 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="ceilometer-central-agent" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.951042 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" containerName="proxy-httpd" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.958890 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.962750 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.963618 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.967454 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:25:52 crc kubenswrapper[4872]: I0126 09:25:52.975463 4872 scope.go:117] "RemoveContainer" containerID="84dab645179878c106b0a586bd291b127e672af3857d1d73b91f24d160a3d99f" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.056233 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-config-data\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.056308 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.056343 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.056372 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-scripts\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.056426 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b378cb-a416-449b-8c1e-0a052aba75bd-run-httpd\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.056445 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b378cb-a416-449b-8c1e-0a052aba75bd-log-httpd\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.056475 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt975\" (UniqueName: \"kubernetes.io/projected/65b378cb-a416-449b-8c1e-0a052aba75bd-kube-api-access-tt975\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.158166 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b378cb-a416-449b-8c1e-0a052aba75bd-run-httpd\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.158222 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b378cb-a416-449b-8c1e-0a052aba75bd-log-httpd\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.158275 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt975\" (UniqueName: \"kubernetes.io/projected/65b378cb-a416-449b-8c1e-0a052aba75bd-kube-api-access-tt975\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.158344 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-config-data\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.158406 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.158437 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.158478 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-scripts\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.159945 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b378cb-a416-449b-8c1e-0a052aba75bd-run-httpd\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.160426 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b378cb-a416-449b-8c1e-0a052aba75bd-log-httpd\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.164281 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-scripts\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.167628 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-config-data\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.168051 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.173480 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.189737 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt975\" (UniqueName: \"kubernetes.io/projected/65b378cb-a416-449b-8c1e-0a052aba75bd-kube-api-access-tt975\") pod \"ceilometer-0\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.207069 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca66fbcd-6af9-42eb-82a1-1e42268194c8" path="/var/lib/kubelet/pods/ca66fbcd-6af9-42eb-82a1-1e42268194c8/volumes" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.219484 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.290528 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.763010 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-srx6m"] Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.773841 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.775919 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.777005 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.779433 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-srx6m"] Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.795779 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.838219 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b378cb-a416-449b-8c1e-0a052aba75bd","Type":"ContainerStarted","Data":"024400034c51e341443484d2b17631ce989e5f9f7c68f352ac1e6bc2c3c1d7c6"} Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.873465 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-srx6m\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.873516 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-scripts\") pod \"nova-cell0-cell-mapping-srx6m\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.873577 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sn8z\" (UniqueName: \"kubernetes.io/projected/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-kube-api-access-8sn8z\") pod \"nova-cell0-cell-mapping-srx6m\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.873633 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-config-data\") pod \"nova-cell0-cell-mapping-srx6m\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.967487 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.969084 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.972853 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.975244 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-config-data\") pod \"nova-cell0-cell-mapping-srx6m\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.975301 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09be6d1-5975-4b6c-9550-e4e329b6ae94-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\") " pod="openstack/nova-scheduler-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.976464 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbwmq\" (UniqueName: \"kubernetes.io/projected/c09be6d1-5975-4b6c-9550-e4e329b6ae94-kube-api-access-jbwmq\") pod \"nova-scheduler-0\" (UID: \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\") " pod="openstack/nova-scheduler-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.976533 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-srx6m\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.976529 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.976627 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-scripts\") pod \"nova-cell0-cell-mapping-srx6m\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.976669 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c09be6d1-5975-4b6c-9550-e4e329b6ae94-config-data\") pod \"nova-scheduler-0\" (UID: \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\") " pod="openstack/nova-scheduler-0" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.976774 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sn8z\" (UniqueName: \"kubernetes.io/projected/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-kube-api-access-8sn8z\") pod \"nova-cell0-cell-mapping-srx6m\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.984751 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-scripts\") pod \"nova-cell0-cell-mapping-srx6m\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.985917 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-config-data\") pod \"nova-cell0-cell-mapping-srx6m\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:53 crc kubenswrapper[4872]: I0126 09:25:53.991139 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-srx6m\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.005304 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sn8z\" (UniqueName: \"kubernetes.io/projected/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-kube-api-access-8sn8z\") pod \"nova-cell0-cell-mapping-srx6m\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.075249 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.077126 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.083119 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.090982 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbwmq\" (UniqueName: \"kubernetes.io/projected/c09be6d1-5975-4b6c-9550-e4e329b6ae94-kube-api-access-jbwmq\") pod \"nova-scheduler-0\" (UID: \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\") " pod="openstack/nova-scheduler-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.091187 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c09be6d1-5975-4b6c-9550-e4e329b6ae94-config-data\") pod \"nova-scheduler-0\" (UID: \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\") " pod="openstack/nova-scheduler-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.091405 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09be6d1-5975-4b6c-9550-e4e329b6ae94-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\") " pod="openstack/nova-scheduler-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.104685 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09be6d1-5975-4b6c-9550-e4e329b6ae94-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\") " pod="openstack/nova-scheduler-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.112658 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c09be6d1-5975-4b6c-9550-e4e329b6ae94-config-data\") pod \"nova-scheduler-0\" (UID: \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\") " pod="openstack/nova-scheduler-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.129568 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbwmq\" (UniqueName: \"kubernetes.io/projected/c09be6d1-5975-4b6c-9550-e4e329b6ae94-kube-api-access-jbwmq\") pod \"nova-scheduler-0\" (UID: \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\") " pod="openstack/nova-scheduler-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.141893 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.168922 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.170903 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.178086 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.182028 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.194086 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-logs\") pod \"nova-api-0\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.194164 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a465b45-2796-49e6-9c8b-e9fb5138052f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.194215 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.194241 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a465b45-2796-49e6-9c8b-e9fb5138052f-config-data\") pod \"nova-metadata-0\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.194262 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-config-data\") pod \"nova-api-0\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.196351 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb8kb\" (UniqueName: \"kubernetes.io/projected/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-kube-api-access-sb8kb\") pod \"nova-api-0\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.196446 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rr62\" (UniqueName: \"kubernetes.io/projected/6a465b45-2796-49e6-9c8b-e9fb5138052f-kube-api-access-9rr62\") pod \"nova-metadata-0\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.196477 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a465b45-2796-49e6-9c8b-e9fb5138052f-logs\") pod \"nova-metadata-0\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.217232 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.299031 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-config-data\") pod \"nova-api-0\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.299160 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb8kb\" (UniqueName: \"kubernetes.io/projected/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-kube-api-access-sb8kb\") pod \"nova-api-0\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.299243 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rr62\" (UniqueName: \"kubernetes.io/projected/6a465b45-2796-49e6-9c8b-e9fb5138052f-kube-api-access-9rr62\") pod \"nova-metadata-0\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.299270 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a465b45-2796-49e6-9c8b-e9fb5138052f-logs\") pod \"nova-metadata-0\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.299357 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-logs\") pod \"nova-api-0\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.299382 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a465b45-2796-49e6-9c8b-e9fb5138052f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.299417 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.299448 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a465b45-2796-49e6-9c8b-e9fb5138052f-config-data\") pod \"nova-metadata-0\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.300331 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a465b45-2796-49e6-9c8b-e9fb5138052f-logs\") pod \"nova-metadata-0\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.300949 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-logs\") pod \"nova-api-0\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.315635 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a465b45-2796-49e6-9c8b-e9fb5138052f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.318088 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.333696 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a465b45-2796-49e6-9c8b-e9fb5138052f-config-data\") pod \"nova-metadata-0\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.343960 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-config-data\") pod \"nova-api-0\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.364401 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb8kb\" (UniqueName: \"kubernetes.io/projected/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-kube-api-access-sb8kb\") pod \"nova-api-0\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.365474 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rr62\" (UniqueName: \"kubernetes.io/projected/6a465b45-2796-49e6-9c8b-e9fb5138052f-kube-api-access-9rr62\") pod \"nova-metadata-0\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.377945 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.379357 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.402392 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w6d7\" (UniqueName: \"kubernetes.io/projected/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-kube-api-access-9w6d7\") pod \"nova-cell1-novncproxy-0\" (UID: \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.402446 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.402532 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.402966 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.414581 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.417618 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.441886 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-gthgc"] Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.476537 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.494268 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.515026 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.515260 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.520138 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w6d7\" (UniqueName: \"kubernetes.io/projected/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-kube-api-access-9w6d7\") pod \"nova-cell1-novncproxy-0\" (UID: \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.549032 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.560964 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w6d7\" (UniqueName: \"kubernetes.io/projected/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-kube-api-access-9w6d7\") pod \"nova-cell1-novncproxy-0\" (UID: \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.625575 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.629054 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.629096 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.629134 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.629202 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-config\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.629236 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q92zr\" (UniqueName: \"kubernetes.io/projected/0855d9b8-26d6-4175-beed-a85fecaa3d29-kube-api-access-q92zr\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.629267 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.629630 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.638430 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.665078 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-gthgc"] Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.695869 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.734903 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-config\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.734962 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q92zr\" (UniqueName: \"kubernetes.io/projected/0855d9b8-26d6-4175-beed-a85fecaa3d29-kube-api-access-q92zr\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.735040 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.735172 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.735187 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.735233 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.737221 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.737829 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.738518 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.739215 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-config\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.747546 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:54 crc kubenswrapper[4872]: I0126 09:25:54.781941 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q92zr\" (UniqueName: \"kubernetes.io/projected/0855d9b8-26d6-4175-beed-a85fecaa3d29-kube-api-access-q92zr\") pod \"dnsmasq-dns-845d6d6f59-gthgc\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.042983 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.208468 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.243281 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-srx6m"] Jan 26 09:25:55 crc kubenswrapper[4872]: W0126 09:25:55.244991 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode23cba8d_bf3b_4683_8f6d_ee2a29db6685.slice/crio-ecacbb0e78bcc5485d80e0e5d9686c5aebc8655afd47053a2e6914a9901adffa WatchSource:0}: Error finding container ecacbb0e78bcc5485d80e0e5d9686c5aebc8655afd47053a2e6914a9901adffa: Status 404 returned error can't find the container with id ecacbb0e78bcc5485d80e0e5d9686c5aebc8655afd47053a2e6914a9901adffa Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.351206 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.420248 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-449lj"] Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.421706 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.424757 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.424996 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.428296 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-449lj"] Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.478865 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.512662 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.557683 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-449lj\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.557769 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-config-data\") pod \"nova-cell1-conductor-db-sync-449lj\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.557858 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj5x8\" (UniqueName: \"kubernetes.io/projected/384d42e1-70a6-44d9-9c4d-9862be13e0ce-kube-api-access-zj5x8\") pod \"nova-cell1-conductor-db-sync-449lj\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.558425 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-scripts\") pod \"nova-cell1-conductor-db-sync-449lj\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.663217 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-449lj\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.663341 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-config-data\") pod \"nova-cell1-conductor-db-sync-449lj\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.663513 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj5x8\" (UniqueName: \"kubernetes.io/projected/384d42e1-70a6-44d9-9c4d-9862be13e0ce-kube-api-access-zj5x8\") pod \"nova-cell1-conductor-db-sync-449lj\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.663612 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-scripts\") pod \"nova-cell1-conductor-db-sync-449lj\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.671265 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-scripts\") pod \"nova-cell1-conductor-db-sync-449lj\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.673921 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-config-data\") pod \"nova-cell1-conductor-db-sync-449lj\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.687822 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-449lj\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.694774 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-gthgc"] Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.701421 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj5x8\" (UniqueName: \"kubernetes.io/projected/384d42e1-70a6-44d9-9c4d-9862be13e0ce-kube-api-access-zj5x8\") pod \"nova-cell1-conductor-db-sync-449lj\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.878324 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b378cb-a416-449b-8c1e-0a052aba75bd","Type":"ContainerStarted","Data":"ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a"} Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.881700 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe","Type":"ContainerStarted","Data":"39c4eb0be1328947c7748abf61ec7eb6aa6b748595ea9d37155a2a0c7308ad36"} Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.885464 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-srx6m" event={"ID":"e23cba8d-bf3b-4683-8f6d-ee2a29db6685","Type":"ContainerStarted","Data":"3c19f590cae21f5eb10846e7021b9c69d79031b1800ae0a641cd0b8e69960457"} Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.885525 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-srx6m" event={"ID":"e23cba8d-bf3b-4683-8f6d-ee2a29db6685","Type":"ContainerStarted","Data":"ecacbb0e78bcc5485d80e0e5d9686c5aebc8655afd47053a2e6914a9901adffa"} Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.892068 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" event={"ID":"0855d9b8-26d6-4175-beed-a85fecaa3d29","Type":"ContainerStarted","Data":"837150f47c0d7e22121774811d91a1ddf01517ab70b9857384633428fc2efe57"} Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.903231 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c09be6d1-5975-4b6c-9550-e4e329b6ae94","Type":"ContainerStarted","Data":"e166ea3d888ba7797ebbbb696053a151d257c8f505efe8815ed0b1e6a060a912"} Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.907684 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6","Type":"ContainerStarted","Data":"7ccbdc1ff908d20b9b7b19f51265ca14f600c56528cd05e86526e1e9994952e5"} Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.908563 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-srx6m" podStartSLOduration=2.9085429449999998 podStartE2EDuration="2.908542945s" podCreationTimestamp="2026-01-26 09:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:25:55.90444398 +0000 UTC m=+1089.213283771" watchObservedRunningTime="2026-01-26 09:25:55.908542945 +0000 UTC m=+1089.217382746" Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.910489 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a465b45-2796-49e6-9c8b-e9fb5138052f","Type":"ContainerStarted","Data":"4752ca35f8afabe530a8b5530c760706e5235150141cb318fcdac77e4761f283"} Jan 26 09:25:55 crc kubenswrapper[4872]: I0126 09:25:55.952580 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:25:56 crc kubenswrapper[4872]: W0126 09:25:56.476756 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod384d42e1_70a6_44d9_9c4d_9862be13e0ce.slice/crio-f7b8f174138143492f2bc998d742325120e1fdfced3f8f2eed84aa2903b43bfe WatchSource:0}: Error finding container f7b8f174138143492f2bc998d742325120e1fdfced3f8f2eed84aa2903b43bfe: Status 404 returned error can't find the container with id f7b8f174138143492f2bc998d742325120e1fdfced3f8f2eed84aa2903b43bfe Jan 26 09:25:56 crc kubenswrapper[4872]: I0126 09:25:56.487592 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-449lj"] Jan 26 09:25:56 crc kubenswrapper[4872]: I0126 09:25:56.927718 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b378cb-a416-449b-8c1e-0a052aba75bd","Type":"ContainerStarted","Data":"8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6"} Jan 26 09:25:56 crc kubenswrapper[4872]: I0126 09:25:56.930242 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-449lj" event={"ID":"384d42e1-70a6-44d9-9c4d-9862be13e0ce","Type":"ContainerStarted","Data":"f7b8f174138143492f2bc998d742325120e1fdfced3f8f2eed84aa2903b43bfe"} Jan 26 09:25:56 crc kubenswrapper[4872]: I0126 09:25:56.933515 4872 generic.go:334] "Generic (PLEG): container finished" podID="0855d9b8-26d6-4175-beed-a85fecaa3d29" containerID="8478f5e2ed375944270c60574080287bacc7b4055e3a7296d102e3d2d1571433" exitCode=0 Jan 26 09:25:56 crc kubenswrapper[4872]: I0126 09:25:56.933641 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" event={"ID":"0855d9b8-26d6-4175-beed-a85fecaa3d29","Type":"ContainerDied","Data":"8478f5e2ed375944270c60574080287bacc7b4055e3a7296d102e3d2d1571433"} Jan 26 09:25:57 crc kubenswrapper[4872]: I0126 09:25:57.772022 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 09:25:57 crc kubenswrapper[4872]: I0126 09:25:57.819009 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:25:57 crc kubenswrapper[4872]: I0126 09:25:57.960059 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-449lj" event={"ID":"384d42e1-70a6-44d9-9c4d-9862be13e0ce","Type":"ContainerStarted","Data":"20e0d61b49f43753f36008fd14e0e79c031c892fc01a6e8baf748d22ba044c71"} Jan 26 09:25:57 crc kubenswrapper[4872]: I0126 09:25:57.968223 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b378cb-a416-449b-8c1e-0a052aba75bd","Type":"ContainerStarted","Data":"175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174"} Jan 26 09:25:57 crc kubenswrapper[4872]: I0126 09:25:57.983830 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" event={"ID":"0855d9b8-26d6-4175-beed-a85fecaa3d29","Type":"ContainerStarted","Data":"ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6"} Jan 26 09:25:57 crc kubenswrapper[4872]: I0126 09:25:57.985135 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:25:57 crc kubenswrapper[4872]: I0126 09:25:57.994586 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-449lj" podStartSLOduration=2.994560493 podStartE2EDuration="2.994560493s" podCreationTimestamp="2026-01-26 09:25:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:25:57.976543771 +0000 UTC m=+1091.285383572" watchObservedRunningTime="2026-01-26 09:25:57.994560493 +0000 UTC m=+1091.303400294" Jan 26 09:25:58 crc kubenswrapper[4872]: I0126 09:25:58.020120 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" podStartSLOduration=4.02009693 podStartE2EDuration="4.02009693s" podCreationTimestamp="2026-01-26 09:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:25:58.007902236 +0000 UTC m=+1091.316742037" watchObservedRunningTime="2026-01-26 09:25:58.02009693 +0000 UTC m=+1091.328936731" Jan 26 09:25:59 crc kubenswrapper[4872]: I0126 09:25:59.591501 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-inspector-0" Jan 26 09:25:59 crc kubenswrapper[4872]: I0126 09:25:59.716513 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ironic-inspector-0" Jan 26 09:25:59 crc kubenswrapper[4872]: I0126 09:25:59.719475 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ironic-inspector-0" Jan 26 09:26:00 crc kubenswrapper[4872]: I0126 09:26:00.030773 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Jan 26 09:26:00 crc kubenswrapper[4872]: I0126 09:26:00.033335 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.044629 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b378cb-a416-449b-8c1e-0a052aba75bd","Type":"ContainerStarted","Data":"0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4"} Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.045899 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.047867 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe","Type":"ContainerStarted","Data":"898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb"} Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.048056 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="14a273ac-90fd-42dd-b5f0-1fd7ca8283fe" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb" gracePeriod=30 Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.053245 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c09be6d1-5975-4b6c-9550-e4e329b6ae94","Type":"ContainerStarted","Data":"7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270"} Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.056132 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6","Type":"ContainerStarted","Data":"894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48"} Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.056166 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6","Type":"ContainerStarted","Data":"ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe"} Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.070156 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.109762407 podStartE2EDuration="10.070129264s" podCreationTimestamp="2026-01-26 09:25:52 +0000 UTC" firstStartedPulling="2026-01-26 09:25:53.800776638 +0000 UTC m=+1087.109616439" lastFinishedPulling="2026-01-26 09:26:00.761143495 +0000 UTC m=+1094.069983296" observedRunningTime="2026-01-26 09:26:02.067027075 +0000 UTC m=+1095.375866906" watchObservedRunningTime="2026-01-26 09:26:02.070129264 +0000 UTC m=+1095.378969085" Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.076429 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6a465b45-2796-49e6-9c8b-e9fb5138052f" containerName="nova-metadata-log" containerID="cri-o://e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80" gracePeriod=30 Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.076985 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a465b45-2796-49e6-9c8b-e9fb5138052f","Type":"ContainerStarted","Data":"8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f"} Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.077069 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a465b45-2796-49e6-9c8b-e9fb5138052f","Type":"ContainerStarted","Data":"e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80"} Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.077012 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6a465b45-2796-49e6-9c8b-e9fb5138052f" containerName="nova-metadata-metadata" containerID="cri-o://8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f" gracePeriod=30 Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.113979 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.758482778 podStartE2EDuration="8.11395695s" podCreationTimestamp="2026-01-26 09:25:54 +0000 UTC" firstStartedPulling="2026-01-26 09:25:55.389398006 +0000 UTC m=+1088.698237807" lastFinishedPulling="2026-01-26 09:26:00.744872178 +0000 UTC m=+1094.053711979" observedRunningTime="2026-01-26 09:26:02.089216624 +0000 UTC m=+1095.398056425" watchObservedRunningTime="2026-01-26 09:26:02.11395695 +0000 UTC m=+1095.422796761" Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.122704 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.581087803 podStartE2EDuration="9.122684424s" podCreationTimestamp="2026-01-26 09:25:53 +0000 UTC" firstStartedPulling="2026-01-26 09:25:55.218589501 +0000 UTC m=+1088.527429292" lastFinishedPulling="2026-01-26 09:26:00.760186112 +0000 UTC m=+1094.069025913" observedRunningTime="2026-01-26 09:26:02.109609387 +0000 UTC m=+1095.418449188" watchObservedRunningTime="2026-01-26 09:26:02.122684424 +0000 UTC m=+1095.431524225" Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.156862 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.893106044 podStartE2EDuration="8.15682222s" podCreationTimestamp="2026-01-26 09:25:54 +0000 UTC" firstStartedPulling="2026-01-26 09:25:55.498523548 +0000 UTC m=+1088.807363349" lastFinishedPulling="2026-01-26 09:26:00.762239724 +0000 UTC m=+1094.071079525" observedRunningTime="2026-01-26 09:26:02.128959194 +0000 UTC m=+1095.437798995" watchObservedRunningTime="2026-01-26 09:26:02.15682222 +0000 UTC m=+1095.465662021" Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.161763 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.87543407 podStartE2EDuration="8.161753877s" podCreationTimestamp="2026-01-26 09:25:54 +0000 UTC" firstStartedPulling="2026-01-26 09:25:55.474275205 +0000 UTC m=+1088.783115006" lastFinishedPulling="2026-01-26 09:26:00.760595012 +0000 UTC m=+1094.069434813" observedRunningTime="2026-01-26 09:26:02.15330482 +0000 UTC m=+1095.462144621" watchObservedRunningTime="2026-01-26 09:26:02.161753877 +0000 UTC m=+1095.470593668" Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.761817 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.933174 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rr62\" (UniqueName: \"kubernetes.io/projected/6a465b45-2796-49e6-9c8b-e9fb5138052f-kube-api-access-9rr62\") pod \"6a465b45-2796-49e6-9c8b-e9fb5138052f\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.933326 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a465b45-2796-49e6-9c8b-e9fb5138052f-config-data\") pod \"6a465b45-2796-49e6-9c8b-e9fb5138052f\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.933432 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a465b45-2796-49e6-9c8b-e9fb5138052f-combined-ca-bundle\") pod \"6a465b45-2796-49e6-9c8b-e9fb5138052f\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.933588 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a465b45-2796-49e6-9c8b-e9fb5138052f-logs\") pod \"6a465b45-2796-49e6-9c8b-e9fb5138052f\" (UID: \"6a465b45-2796-49e6-9c8b-e9fb5138052f\") " Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.934240 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a465b45-2796-49e6-9c8b-e9fb5138052f-logs" (OuterVolumeSpecName: "logs") pod "6a465b45-2796-49e6-9c8b-e9fb5138052f" (UID: "6a465b45-2796-49e6-9c8b-e9fb5138052f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.945322 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a465b45-2796-49e6-9c8b-e9fb5138052f-kube-api-access-9rr62" (OuterVolumeSpecName: "kube-api-access-9rr62") pod "6a465b45-2796-49e6-9c8b-e9fb5138052f" (UID: "6a465b45-2796-49e6-9c8b-e9fb5138052f"). InnerVolumeSpecName "kube-api-access-9rr62". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.967419 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a465b45-2796-49e6-9c8b-e9fb5138052f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a465b45-2796-49e6-9c8b-e9fb5138052f" (UID: "6a465b45-2796-49e6-9c8b-e9fb5138052f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:02 crc kubenswrapper[4872]: I0126 09:26:02.969979 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a465b45-2796-49e6-9c8b-e9fb5138052f-config-data" (OuterVolumeSpecName: "config-data") pod "6a465b45-2796-49e6-9c8b-e9fb5138052f" (UID: "6a465b45-2796-49e6-9c8b-e9fb5138052f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.037030 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a465b45-2796-49e6-9c8b-e9fb5138052f-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.037092 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a465b45-2796-49e6-9c8b-e9fb5138052f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.037118 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6a465b45-2796-49e6-9c8b-e9fb5138052f-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.037135 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rr62\" (UniqueName: \"kubernetes.io/projected/6a465b45-2796-49e6-9c8b-e9fb5138052f-kube-api-access-9rr62\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.090396 4872 generic.go:334] "Generic (PLEG): container finished" podID="6a465b45-2796-49e6-9c8b-e9fb5138052f" containerID="8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f" exitCode=0 Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.090456 4872 generic.go:334] "Generic (PLEG): container finished" podID="6a465b45-2796-49e6-9c8b-e9fb5138052f" containerID="e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80" exitCode=143 Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.090566 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a465b45-2796-49e6-9c8b-e9fb5138052f","Type":"ContainerDied","Data":"8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f"} Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.090790 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a465b45-2796-49e6-9c8b-e9fb5138052f","Type":"ContainerDied","Data":"e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80"} Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.090861 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6a465b45-2796-49e6-9c8b-e9fb5138052f","Type":"ContainerDied","Data":"4752ca35f8afabe530a8b5530c760706e5235150141cb318fcdac77e4761f283"} Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.090907 4872 scope.go:117] "RemoveContainer" containerID="8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.090663 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.129245 4872 scope.go:117] "RemoveContainer" containerID="e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.145045 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.160991 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.173752 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:03 crc kubenswrapper[4872]: E0126 09:26:03.174525 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a465b45-2796-49e6-9c8b-e9fb5138052f" containerName="nova-metadata-metadata" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.174545 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a465b45-2796-49e6-9c8b-e9fb5138052f" containerName="nova-metadata-metadata" Jan 26 09:26:03 crc kubenswrapper[4872]: E0126 09:26:03.174554 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a465b45-2796-49e6-9c8b-e9fb5138052f" containerName="nova-metadata-log" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.174562 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a465b45-2796-49e6-9c8b-e9fb5138052f" containerName="nova-metadata-log" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.174940 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a465b45-2796-49e6-9c8b-e9fb5138052f" containerName="nova-metadata-log" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.174967 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a465b45-2796-49e6-9c8b-e9fb5138052f" containerName="nova-metadata-metadata" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.176453 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.181532 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.181943 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.214939 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a465b45-2796-49e6-9c8b-e9fb5138052f" path="/var/lib/kubelet/pods/6a465b45-2796-49e6-9c8b-e9fb5138052f/volumes" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.218874 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.237133 4872 scope.go:117] "RemoveContainer" containerID="8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f" Jan 26 09:26:03 crc kubenswrapper[4872]: E0126 09:26:03.237563 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f\": container with ID starting with 8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f not found: ID does not exist" containerID="8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.237604 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f"} err="failed to get container status \"8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f\": rpc error: code = NotFound desc = could not find container \"8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f\": container with ID starting with 8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f not found: ID does not exist" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.237630 4872 scope.go:117] "RemoveContainer" containerID="e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80" Jan 26 09:26:03 crc kubenswrapper[4872]: E0126 09:26:03.237818 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80\": container with ID starting with e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80 not found: ID does not exist" containerID="e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.237842 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80"} err="failed to get container status \"e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80\": rpc error: code = NotFound desc = could not find container \"e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80\": container with ID starting with e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80 not found: ID does not exist" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.237856 4872 scope.go:117] "RemoveContainer" containerID="8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.238207 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f"} err="failed to get container status \"8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f\": rpc error: code = NotFound desc = could not find container \"8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f\": container with ID starting with 8cf8d98f5dab45c96ad71b0fbc44838d2d7382baa6cbfbc408f89c62e3476d3f not found: ID does not exist" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.238337 4872 scope.go:117] "RemoveContainer" containerID="e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.238502 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80"} err="failed to get container status \"e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80\": rpc error: code = NotFound desc = could not find container \"e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80\": container with ID starting with e73c5fb3ca9f40b3743abaf280c3b08cf7f06a3822d61faa1461cd05ff90aa80 not found: ID does not exist" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.351816 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-config-data\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.351978 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.352039 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.352072 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bd8b\" (UniqueName: \"kubernetes.io/projected/62d627e7-d04f-4673-a7c4-3d38488f0244-kube-api-access-8bd8b\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.352163 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62d627e7-d04f-4673-a7c4-3d38488f0244-logs\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.454149 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.454216 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.454259 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bd8b\" (UniqueName: \"kubernetes.io/projected/62d627e7-d04f-4673-a7c4-3d38488f0244-kube-api-access-8bd8b\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.454338 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62d627e7-d04f-4673-a7c4-3d38488f0244-logs\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.454423 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-config-data\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.454939 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62d627e7-d04f-4673-a7c4-3d38488f0244-logs\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.460158 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.461740 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-config-data\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.470486 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.476123 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bd8b\" (UniqueName: \"kubernetes.io/projected/62d627e7-d04f-4673-a7c4-3d38488f0244-kube-api-access-8bd8b\") pod \"nova-metadata-0\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " pod="openstack/nova-metadata-0" Jan 26 09:26:03 crc kubenswrapper[4872]: I0126 09:26:03.519235 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:26:04 crc kubenswrapper[4872]: I0126 09:26:04.030069 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:04 crc kubenswrapper[4872]: W0126 09:26:04.057060 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62d627e7_d04f_4673_a7c4_3d38488f0244.slice/crio-ac6575b541dc40de27697bc0cd95db5c4b9e4a1452a94b098c879fd2c7f3887e WatchSource:0}: Error finding container ac6575b541dc40de27697bc0cd95db5c4b9e4a1452a94b098c879fd2c7f3887e: Status 404 returned error can't find the container with id ac6575b541dc40de27697bc0cd95db5c4b9e4a1452a94b098c879fd2c7f3887e Jan 26 09:26:04 crc kubenswrapper[4872]: I0126 09:26:04.111252 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62d627e7-d04f-4673-a7c4-3d38488f0244","Type":"ContainerStarted","Data":"ac6575b541dc40de27697bc0cd95db5c4b9e4a1452a94b098c879fd2c7f3887e"} Jan 26 09:26:04 crc kubenswrapper[4872]: I0126 09:26:04.403625 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 26 09:26:04 crc kubenswrapper[4872]: I0126 09:26:04.404220 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 26 09:26:04 crc kubenswrapper[4872]: I0126 09:26:04.452861 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 26 09:26:04 crc kubenswrapper[4872]: I0126 09:26:04.479264 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 09:26:04 crc kubenswrapper[4872]: I0126 09:26:04.479371 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 09:26:04 crc kubenswrapper[4872]: I0126 09:26:04.697163 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.045028 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.118924 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qnrbz"] Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.119216 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" podUID="99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" containerName="dnsmasq-dns" containerID="cri-o://4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d" gracePeriod=10 Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.174792 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62d627e7-d04f-4673-a7c4-3d38488f0244","Type":"ContainerStarted","Data":"7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2"} Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.174879 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62d627e7-d04f-4673-a7c4-3d38488f0244","Type":"ContainerStarted","Data":"45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61"} Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.255543 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.277441 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.277410912 podStartE2EDuration="2.277410912s" podCreationTimestamp="2026-01-26 09:26:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:26:05.203618297 +0000 UTC m=+1098.512458098" watchObservedRunningTime="2026-01-26 09:26:05.277410912 +0000 UTC m=+1098.586250713" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.572079 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.572182 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.680331 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.852239 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grtb4\" (UniqueName: \"kubernetes.io/projected/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-kube-api-access-grtb4\") pod \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.852378 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-dns-svc\") pod \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.852457 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-config\") pod \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.852481 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-ovsdbserver-nb\") pod \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.852500 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-dns-swift-storage-0\") pod \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.852523 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-ovsdbserver-sb\") pod \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\" (UID: \"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389\") " Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.858875 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-kube-api-access-grtb4" (OuterVolumeSpecName: "kube-api-access-grtb4") pod "99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" (UID: "99c1bb1d-75c2-4a9e-98e6-d8c80fe19389"). InnerVolumeSpecName "kube-api-access-grtb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.903319 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" (UID: "99c1bb1d-75c2-4a9e-98e6-d8c80fe19389"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.930214 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" (UID: "99c1bb1d-75c2-4a9e-98e6-d8c80fe19389"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.944824 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-config" (OuterVolumeSpecName: "config") pod "99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" (UID: "99c1bb1d-75c2-4a9e-98e6-d8c80fe19389"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.948196 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" (UID: "99c1bb1d-75c2-4a9e-98e6-d8c80fe19389"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.955748 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grtb4\" (UniqueName: \"kubernetes.io/projected/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-kube-api-access-grtb4\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.955783 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.955823 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.955838 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.955850 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:05 crc kubenswrapper[4872]: I0126 09:26:05.988319 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" (UID: "99c1bb1d-75c2-4a9e-98e6-d8c80fe19389"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.058238 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.186937 4872 generic.go:334] "Generic (PLEG): container finished" podID="99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" containerID="4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d" exitCode=0 Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.187226 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" event={"ID":"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389","Type":"ContainerDied","Data":"4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d"} Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.187251 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.187293 4872 scope.go:117] "RemoveContainer" containerID="4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d" Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.187274 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-qnrbz" event={"ID":"99c1bb1d-75c2-4a9e-98e6-d8c80fe19389","Type":"ContainerDied","Data":"fb4000f3f794ce5f11cfbb75b7107a420ce00419f0e3e7bc18664f5e0d1889bd"} Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.190498 4872 generic.go:334] "Generic (PLEG): container finished" podID="e23cba8d-bf3b-4683-8f6d-ee2a29db6685" containerID="3c19f590cae21f5eb10846e7021b9c69d79031b1800ae0a641cd0b8e69960457" exitCode=0 Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.190568 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-srx6m" event={"ID":"e23cba8d-bf3b-4683-8f6d-ee2a29db6685","Type":"ContainerDied","Data":"3c19f590cae21f5eb10846e7021b9c69d79031b1800ae0a641cd0b8e69960457"} Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.223530 4872 scope.go:117] "RemoveContainer" containerID="b8dafe34e2d01960c568c047f2f3f07cb31abbcbca4a0fd41efbeb28bb54a2de" Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.255605 4872 scope.go:117] "RemoveContainer" containerID="4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d" Jan 26 09:26:06 crc kubenswrapper[4872]: E0126 09:26:06.256664 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d\": container with ID starting with 4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d not found: ID does not exist" containerID="4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d" Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.256841 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d"} err="failed to get container status \"4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d\": rpc error: code = NotFound desc = could not find container \"4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d\": container with ID starting with 4a57e59c74d30f38df7d5a952e5cbbc54dc9621076a2e8282008e4ca3bd1cc1d not found: ID does not exist" Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.256969 4872 scope.go:117] "RemoveContainer" containerID="b8dafe34e2d01960c568c047f2f3f07cb31abbcbca4a0fd41efbeb28bb54a2de" Jan 26 09:26:06 crc kubenswrapper[4872]: E0126 09:26:06.257425 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8dafe34e2d01960c568c047f2f3f07cb31abbcbca4a0fd41efbeb28bb54a2de\": container with ID starting with b8dafe34e2d01960c568c047f2f3f07cb31abbcbca4a0fd41efbeb28bb54a2de not found: ID does not exist" containerID="b8dafe34e2d01960c568c047f2f3f07cb31abbcbca4a0fd41efbeb28bb54a2de" Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.257521 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8dafe34e2d01960c568c047f2f3f07cb31abbcbca4a0fd41efbeb28bb54a2de"} err="failed to get container status \"b8dafe34e2d01960c568c047f2f3f07cb31abbcbca4a0fd41efbeb28bb54a2de\": rpc error: code = NotFound desc = could not find container \"b8dafe34e2d01960c568c047f2f3f07cb31abbcbca4a0fd41efbeb28bb54a2de\": container with ID starting with b8dafe34e2d01960c568c047f2f3f07cb31abbcbca4a0fd41efbeb28bb54a2de not found: ID does not exist" Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.273088 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qnrbz"] Jan 26 09:26:06 crc kubenswrapper[4872]: I0126 09:26:06.282180 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-qnrbz"] Jan 26 09:26:07 crc kubenswrapper[4872]: I0126 09:26:07.214891 4872 generic.go:334] "Generic (PLEG): container finished" podID="384d42e1-70a6-44d9-9c4d-9862be13e0ce" containerID="20e0d61b49f43753f36008fd14e0e79c031c892fc01a6e8baf748d22ba044c71" exitCode=0 Jan 26 09:26:07 crc kubenswrapper[4872]: I0126 09:26:07.221672 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" path="/var/lib/kubelet/pods/99c1bb1d-75c2-4a9e-98e6-d8c80fe19389/volumes" Jan 26 09:26:07 crc kubenswrapper[4872]: I0126 09:26:07.222457 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-449lj" event={"ID":"384d42e1-70a6-44d9-9c4d-9862be13e0ce","Type":"ContainerDied","Data":"20e0d61b49f43753f36008fd14e0e79c031c892fc01a6e8baf748d22ba044c71"} Jan 26 09:26:07 crc kubenswrapper[4872]: I0126 09:26:07.704767 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:26:07 crc kubenswrapper[4872]: I0126 09:26:07.906024 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-config-data\") pod \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " Jan 26 09:26:07 crc kubenswrapper[4872]: I0126 09:26:07.906450 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-combined-ca-bundle\") pod \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " Jan 26 09:26:07 crc kubenswrapper[4872]: I0126 09:26:07.906554 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-scripts\") pod \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " Jan 26 09:26:07 crc kubenswrapper[4872]: I0126 09:26:07.906597 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sn8z\" (UniqueName: \"kubernetes.io/projected/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-kube-api-access-8sn8z\") pod \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\" (UID: \"e23cba8d-bf3b-4683-8f6d-ee2a29db6685\") " Jan 26 09:26:07 crc kubenswrapper[4872]: I0126 09:26:07.915678 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-kube-api-access-8sn8z" (OuterVolumeSpecName: "kube-api-access-8sn8z") pod "e23cba8d-bf3b-4683-8f6d-ee2a29db6685" (UID: "e23cba8d-bf3b-4683-8f6d-ee2a29db6685"). InnerVolumeSpecName "kube-api-access-8sn8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:07 crc kubenswrapper[4872]: I0126 09:26:07.917119 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-scripts" (OuterVolumeSpecName: "scripts") pod "e23cba8d-bf3b-4683-8f6d-ee2a29db6685" (UID: "e23cba8d-bf3b-4683-8f6d-ee2a29db6685"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:07 crc kubenswrapper[4872]: I0126 09:26:07.949571 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e23cba8d-bf3b-4683-8f6d-ee2a29db6685" (UID: "e23cba8d-bf3b-4683-8f6d-ee2a29db6685"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:07 crc kubenswrapper[4872]: I0126 09:26:07.951267 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-config-data" (OuterVolumeSpecName: "config-data") pod "e23cba8d-bf3b-4683-8f6d-ee2a29db6685" (UID: "e23cba8d-bf3b-4683-8f6d-ee2a29db6685"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.012395 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.012621 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.012790 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.013008 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sn8z\" (UniqueName: \"kubernetes.io/projected/e23cba8d-bf3b-4683-8f6d-ee2a29db6685-kube-api-access-8sn8z\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.227920 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-srx6m" event={"ID":"e23cba8d-bf3b-4683-8f6d-ee2a29db6685","Type":"ContainerDied","Data":"ecacbb0e78bcc5485d80e0e5d9686c5aebc8655afd47053a2e6914a9901adffa"} Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.228028 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecacbb0e78bcc5485d80e0e5d9686c5aebc8655afd47053a2e6914a9901adffa" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.227977 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-srx6m" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.417165 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.417493 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" containerName="nova-api-log" containerID="cri-o://ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe" gracePeriod=30 Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.417897 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" containerName="nova-api-api" containerID="cri-o://894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48" gracePeriod=30 Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.429339 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.429678 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c09be6d1-5975-4b6c-9550-e4e329b6ae94" containerName="nova-scheduler-scheduler" containerID="cri-o://7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270" gracePeriod=30 Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.511482 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.512392 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="62d627e7-d04f-4673-a7c4-3d38488f0244" containerName="nova-metadata-log" containerID="cri-o://45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61" gracePeriod=30 Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.512575 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="62d627e7-d04f-4673-a7c4-3d38488f0244" containerName="nova-metadata-metadata" containerID="cri-o://7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2" gracePeriod=30 Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.521032 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.521096 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.687992 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.728944 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-combined-ca-bundle\") pod \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.729220 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zj5x8\" (UniqueName: \"kubernetes.io/projected/384d42e1-70a6-44d9-9c4d-9862be13e0ce-kube-api-access-zj5x8\") pod \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.729383 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-scripts\") pod \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.729475 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-config-data\") pod \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\" (UID: \"384d42e1-70a6-44d9-9c4d-9862be13e0ce\") " Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.739972 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-scripts" (OuterVolumeSpecName: "scripts") pod "384d42e1-70a6-44d9-9c4d-9862be13e0ce" (UID: "384d42e1-70a6-44d9-9c4d-9862be13e0ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.740151 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/384d42e1-70a6-44d9-9c4d-9862be13e0ce-kube-api-access-zj5x8" (OuterVolumeSpecName: "kube-api-access-zj5x8") pod "384d42e1-70a6-44d9-9c4d-9862be13e0ce" (UID: "384d42e1-70a6-44d9-9c4d-9862be13e0ce"). InnerVolumeSpecName "kube-api-access-zj5x8". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.763217 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "384d42e1-70a6-44d9-9c4d-9862be13e0ce" (UID: "384d42e1-70a6-44d9-9c4d-9862be13e0ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.773538 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-config-data" (OuterVolumeSpecName: "config-data") pod "384d42e1-70a6-44d9-9c4d-9862be13e0ce" (UID: "384d42e1-70a6-44d9-9c4d-9862be13e0ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.832947 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.832995 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zj5x8\" (UniqueName: \"kubernetes.io/projected/384d42e1-70a6-44d9-9c4d-9862be13e0ce-kube-api-access-zj5x8\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.833012 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:08 crc kubenswrapper[4872]: I0126 09:26:08.833024 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/384d42e1-70a6-44d9-9c4d-9862be13e0ce-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.239449 4872 generic.go:334] "Generic (PLEG): container finished" podID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" containerID="ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe" exitCode=143 Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.239610 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6","Type":"ContainerDied","Data":"ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe"} Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.245262 4872 generic.go:334] "Generic (PLEG): container finished" podID="62d627e7-d04f-4673-a7c4-3d38488f0244" containerID="45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61" exitCode=143 Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.245318 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62d627e7-d04f-4673-a7c4-3d38488f0244","Type":"ContainerDied","Data":"45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61"} Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.247872 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-449lj" event={"ID":"384d42e1-70a6-44d9-9c4d-9862be13e0ce","Type":"ContainerDied","Data":"f7b8f174138143492f2bc998d742325120e1fdfced3f8f2eed84aa2903b43bfe"} Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.247910 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7b8f174138143492f2bc998d742325120e1fdfced3f8f2eed84aa2903b43bfe" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.247944 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-449lj" Jan 26 09:26:09 crc kubenswrapper[4872]: E0126 09:26:09.407555 4872 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 26 09:26:09 crc kubenswrapper[4872]: E0126 09:26:09.409256 4872 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 26 09:26:09 crc kubenswrapper[4872]: E0126 09:26:09.411076 4872 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Jan 26 09:26:09 crc kubenswrapper[4872]: E0126 09:26:09.411118 4872 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c09be6d1-5975-4b6c-9550-e4e329b6ae94" containerName="nova-scheduler-scheduler" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.569858 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 26 09:26:09 crc kubenswrapper[4872]: E0126 09:26:09.570715 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e23cba8d-bf3b-4683-8f6d-ee2a29db6685" containerName="nova-manage" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.570737 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e23cba8d-bf3b-4683-8f6d-ee2a29db6685" containerName="nova-manage" Jan 26 09:26:09 crc kubenswrapper[4872]: E0126 09:26:09.570745 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="384d42e1-70a6-44d9-9c4d-9862be13e0ce" containerName="nova-cell1-conductor-db-sync" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.570752 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="384d42e1-70a6-44d9-9c4d-9862be13e0ce" containerName="nova-cell1-conductor-db-sync" Jan 26 09:26:09 crc kubenswrapper[4872]: E0126 09:26:09.570781 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" containerName="init" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.570788 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" containerName="init" Jan 26 09:26:09 crc kubenswrapper[4872]: E0126 09:26:09.570868 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" containerName="dnsmasq-dns" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.570876 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" containerName="dnsmasq-dns" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.571076 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e23cba8d-bf3b-4683-8f6d-ee2a29db6685" containerName="nova-manage" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.571090 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="99c1bb1d-75c2-4a9e-98e6-d8c80fe19389" containerName="dnsmasq-dns" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.571100 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="384d42e1-70a6-44d9-9c4d-9862be13e0ce" containerName="nova-cell1-conductor-db-sync" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.571901 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.577433 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.590830 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.649646 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5d286f-e785-4a4b-b96e-392ea79d7d2e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"dc5d286f-e785-4a4b-b96e-392ea79d7d2e\") " pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.649758 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc5d286f-e785-4a4b-b96e-392ea79d7d2e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"dc5d286f-e785-4a4b-b96e-392ea79d7d2e\") " pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.649826 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kwgl\" (UniqueName: \"kubernetes.io/projected/dc5d286f-e785-4a4b-b96e-392ea79d7d2e-kube-api-access-8kwgl\") pod \"nova-cell1-conductor-0\" (UID: \"dc5d286f-e785-4a4b-b96e-392ea79d7d2e\") " pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.752137 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5d286f-e785-4a4b-b96e-392ea79d7d2e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"dc5d286f-e785-4a4b-b96e-392ea79d7d2e\") " pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.752859 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc5d286f-e785-4a4b-b96e-392ea79d7d2e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"dc5d286f-e785-4a4b-b96e-392ea79d7d2e\") " pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.753520 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kwgl\" (UniqueName: \"kubernetes.io/projected/dc5d286f-e785-4a4b-b96e-392ea79d7d2e-kube-api-access-8kwgl\") pod \"nova-cell1-conductor-0\" (UID: \"dc5d286f-e785-4a4b-b96e-392ea79d7d2e\") " pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.759513 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc5d286f-e785-4a4b-b96e-392ea79d7d2e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"dc5d286f-e785-4a4b-b96e-392ea79d7d2e\") " pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.774976 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc5d286f-e785-4a4b-b96e-392ea79d7d2e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"dc5d286f-e785-4a4b-b96e-392ea79d7d2e\") " pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.777209 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kwgl\" (UniqueName: \"kubernetes.io/projected/dc5d286f-e785-4a4b-b96e-392ea79d7d2e-kube-api-access-8kwgl\") pod \"nova-cell1-conductor-0\" (UID: \"dc5d286f-e785-4a4b-b96e-392ea79d7d2e\") " pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:09 crc kubenswrapper[4872]: I0126 09:26:09.983297 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.006915 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.060659 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c09be6d1-5975-4b6c-9550-e4e329b6ae94-config-data\") pod \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\" (UID: \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\") " Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.060757 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09be6d1-5975-4b6c-9550-e4e329b6ae94-combined-ca-bundle\") pod \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\" (UID: \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\") " Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.060874 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbwmq\" (UniqueName: \"kubernetes.io/projected/c09be6d1-5975-4b6c-9550-e4e329b6ae94-kube-api-access-jbwmq\") pod \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\" (UID: \"c09be6d1-5975-4b6c-9550-e4e329b6ae94\") " Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.077051 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c09be6d1-5975-4b6c-9550-e4e329b6ae94-kube-api-access-jbwmq" (OuterVolumeSpecName: "kube-api-access-jbwmq") pod "c09be6d1-5975-4b6c-9550-e4e329b6ae94" (UID: "c09be6d1-5975-4b6c-9550-e4e329b6ae94"). InnerVolumeSpecName "kube-api-access-jbwmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.099773 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c09be6d1-5975-4b6c-9550-e4e329b6ae94-config-data" (OuterVolumeSpecName: "config-data") pod "c09be6d1-5975-4b6c-9550-e4e329b6ae94" (UID: "c09be6d1-5975-4b6c-9550-e4e329b6ae94"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.105089 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c09be6d1-5975-4b6c-9550-e4e329b6ae94-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c09be6d1-5975-4b6c-9550-e4e329b6ae94" (UID: "c09be6d1-5975-4b6c-9550-e4e329b6ae94"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.164686 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c09be6d1-5975-4b6c-9550-e4e329b6ae94-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.164720 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09be6d1-5975-4b6c-9550-e4e329b6ae94-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.164735 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbwmq\" (UniqueName: \"kubernetes.io/projected/c09be6d1-5975-4b6c-9550-e4e329b6ae94-kube-api-access-jbwmq\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.166528 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.262392 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.262287 4872 generic.go:334] "Generic (PLEG): container finished" podID="c09be6d1-5975-4b6c-9550-e4e329b6ae94" containerID="7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270" exitCode=0 Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.262941 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c09be6d1-5975-4b6c-9550-e4e329b6ae94","Type":"ContainerDied","Data":"7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270"} Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.263263 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c09be6d1-5975-4b6c-9550-e4e329b6ae94","Type":"ContainerDied","Data":"e166ea3d888ba7797ebbbb696053a151d257c8f505efe8815ed0b1e6a060a912"} Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.263327 4872 scope.go:117] "RemoveContainer" containerID="7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.269491 4872 generic.go:334] "Generic (PLEG): container finished" podID="62d627e7-d04f-4673-a7c4-3d38488f0244" containerID="7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2" exitCode=0 Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.269552 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62d627e7-d04f-4673-a7c4-3d38488f0244","Type":"ContainerDied","Data":"7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2"} Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.269591 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"62d627e7-d04f-4673-a7c4-3d38488f0244","Type":"ContainerDied","Data":"ac6575b541dc40de27697bc0cd95db5c4b9e4a1452a94b098c879fd2c7f3887e"} Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.269670 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.303347 4872 scope.go:117] "RemoveContainer" containerID="7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270" Jan 26 09:26:10 crc kubenswrapper[4872]: E0126 09:26:10.304677 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270\": container with ID starting with 7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270 not found: ID does not exist" containerID="7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.304740 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270"} err="failed to get container status \"7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270\": rpc error: code = NotFound desc = could not find container \"7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270\": container with ID starting with 7422102d0074fc3c8628b15a0e56c6e29cf91502be327acc729015d684b08270 not found: ID does not exist" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.304776 4872 scope.go:117] "RemoveContainer" containerID="7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.324834 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.347365 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.369026 4872 scope.go:117] "RemoveContainer" containerID="45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.370773 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-combined-ca-bundle\") pod \"62d627e7-d04f-4673-a7c4-3d38488f0244\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.370896 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62d627e7-d04f-4673-a7c4-3d38488f0244-logs\") pod \"62d627e7-d04f-4673-a7c4-3d38488f0244\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.371147 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bd8b\" (UniqueName: \"kubernetes.io/projected/62d627e7-d04f-4673-a7c4-3d38488f0244-kube-api-access-8bd8b\") pod \"62d627e7-d04f-4673-a7c4-3d38488f0244\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.371195 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-config-data\") pod \"62d627e7-d04f-4673-a7c4-3d38488f0244\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.371267 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-nova-metadata-tls-certs\") pod \"62d627e7-d04f-4673-a7c4-3d38488f0244\" (UID: \"62d627e7-d04f-4673-a7c4-3d38488f0244\") " Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.372325 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62d627e7-d04f-4673-a7c4-3d38488f0244-logs" (OuterVolumeSpecName: "logs") pod "62d627e7-d04f-4673-a7c4-3d38488f0244" (UID: "62d627e7-d04f-4673-a7c4-3d38488f0244"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.374335 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62d627e7-d04f-4673-a7c4-3d38488f0244-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.381909 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:26:10 crc kubenswrapper[4872]: E0126 09:26:10.386034 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d627e7-d04f-4673-a7c4-3d38488f0244" containerName="nova-metadata-metadata" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.386062 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d627e7-d04f-4673-a7c4-3d38488f0244" containerName="nova-metadata-metadata" Jan 26 09:26:10 crc kubenswrapper[4872]: E0126 09:26:10.386103 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62d627e7-d04f-4673-a7c4-3d38488f0244" containerName="nova-metadata-log" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.386113 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="62d627e7-d04f-4673-a7c4-3d38488f0244" containerName="nova-metadata-log" Jan 26 09:26:10 crc kubenswrapper[4872]: E0126 09:26:10.386149 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c09be6d1-5975-4b6c-9550-e4e329b6ae94" containerName="nova-scheduler-scheduler" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.386155 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c09be6d1-5975-4b6c-9550-e4e329b6ae94" containerName="nova-scheduler-scheduler" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.386497 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d627e7-d04f-4673-a7c4-3d38488f0244" containerName="nova-metadata-log" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.386520 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="62d627e7-d04f-4673-a7c4-3d38488f0244" containerName="nova-metadata-metadata" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.386545 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="c09be6d1-5975-4b6c-9550-e4e329b6ae94" containerName="nova-scheduler-scheduler" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.387891 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.382497 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62d627e7-d04f-4673-a7c4-3d38488f0244-kube-api-access-8bd8b" (OuterVolumeSpecName: "kube-api-access-8bd8b") pod "62d627e7-d04f-4673-a7c4-3d38488f0244" (UID: "62d627e7-d04f-4673-a7c4-3d38488f0244"). InnerVolumeSpecName "kube-api-access-8bd8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.410184 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.416016 4872 scope.go:117] "RemoveContainer" containerID="7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2" Jan 26 09:26:10 crc kubenswrapper[4872]: E0126 09:26:10.417812 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2\": container with ID starting with 7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2 not found: ID does not exist" containerID="7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.417932 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2"} err="failed to get container status \"7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2\": rpc error: code = NotFound desc = could not find container \"7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2\": container with ID starting with 7a7f59406ce0ee0d75187201faa1cbe878b15871451e6e1434e26287ee8f7ea2 not found: ID does not exist" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.417964 4872 scope.go:117] "RemoveContainer" containerID="45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61" Jan 26 09:26:10 crc kubenswrapper[4872]: E0126 09:26:10.418882 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61\": container with ID starting with 45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61 not found: ID does not exist" containerID="45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.418987 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61"} err="failed to get container status \"45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61\": rpc error: code = NotFound desc = could not find container \"45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61\": container with ID starting with 45d589d862095f2b1180b79d056a307551db5acc13e80fafd5b207e79870ea61 not found: ID does not exist" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.421689 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62d627e7-d04f-4673-a7c4-3d38488f0244" (UID: "62d627e7-d04f-4673-a7c4-3d38488f0244"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.426968 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.432915 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "62d627e7-d04f-4673-a7c4-3d38488f0244" (UID: "62d627e7-d04f-4673-a7c4-3d38488f0244"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.441214 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-config-data" (OuterVolumeSpecName: "config-data") pod "62d627e7-d04f-4673-a7c4-3d38488f0244" (UID: "62d627e7-d04f-4673-a7c4-3d38488f0244"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.475364 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19b7b80d-07f5-472b-b3df-dce1de4223a6-config-data\") pod \"nova-scheduler-0\" (UID: \"19b7b80d-07f5-472b-b3df-dce1de4223a6\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.475743 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnc69\" (UniqueName: \"kubernetes.io/projected/19b7b80d-07f5-472b-b3df-dce1de4223a6-kube-api-access-hnc69\") pod \"nova-scheduler-0\" (UID: \"19b7b80d-07f5-472b-b3df-dce1de4223a6\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.475935 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b7b80d-07f5-472b-b3df-dce1de4223a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"19b7b80d-07f5-472b-b3df-dce1de4223a6\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.476046 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.476105 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bd8b\" (UniqueName: \"kubernetes.io/projected/62d627e7-d04f-4673-a7c4-3d38488f0244-kube-api-access-8bd8b\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.476162 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.476226 4872 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/62d627e7-d04f-4673-a7c4-3d38488f0244-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:10 crc kubenswrapper[4872]: W0126 09:26:10.497981 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc5d286f_e785_4a4b_b96e_392ea79d7d2e.slice/crio-ca0049935f85e434781633e8e54495946c182e25f9ccc62d6a62643bc15b5414 WatchSource:0}: Error finding container ca0049935f85e434781633e8e54495946c182e25f9ccc62d6a62643bc15b5414: Status 404 returned error can't find the container with id ca0049935f85e434781633e8e54495946c182e25f9ccc62d6a62643bc15b5414 Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.501361 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.576863 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnc69\" (UniqueName: \"kubernetes.io/projected/19b7b80d-07f5-472b-b3df-dce1de4223a6-kube-api-access-hnc69\") pod \"nova-scheduler-0\" (UID: \"19b7b80d-07f5-472b-b3df-dce1de4223a6\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.576988 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b7b80d-07f5-472b-b3df-dce1de4223a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"19b7b80d-07f5-472b-b3df-dce1de4223a6\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.577014 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19b7b80d-07f5-472b-b3df-dce1de4223a6-config-data\") pod \"nova-scheduler-0\" (UID: \"19b7b80d-07f5-472b-b3df-dce1de4223a6\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.580915 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19b7b80d-07f5-472b-b3df-dce1de4223a6-config-data\") pod \"nova-scheduler-0\" (UID: \"19b7b80d-07f5-472b-b3df-dce1de4223a6\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.581389 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b7b80d-07f5-472b-b3df-dce1de4223a6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"19b7b80d-07f5-472b-b3df-dce1de4223a6\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.612401 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnc69\" (UniqueName: \"kubernetes.io/projected/19b7b80d-07f5-472b-b3df-dce1de4223a6-kube-api-access-hnc69\") pod \"nova-scheduler-0\" (UID: \"19b7b80d-07f5-472b-b3df-dce1de4223a6\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.628482 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.645605 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.674926 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.677250 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.678433 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-config-data\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.678479 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4c81032-5e90-4b7d-a27d-a46b7e052269-logs\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.678536 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4vr6\" (UniqueName: \"kubernetes.io/projected/a4c81032-5e90-4b7d-a27d-a46b7e052269-kube-api-access-t4vr6\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.678564 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.678600 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.679966 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.680090 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.685042 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.714867 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.784440 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.784970 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.785232 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-config-data\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.785376 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4c81032-5e90-4b7d-a27d-a46b7e052269-logs\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.785597 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4vr6\" (UniqueName: \"kubernetes.io/projected/a4c81032-5e90-4b7d-a27d-a46b7e052269-kube-api-access-t4vr6\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.786461 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4c81032-5e90-4b7d-a27d-a46b7e052269-logs\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.790656 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-config-data\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.793909 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.797221 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.808228 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4vr6\" (UniqueName: \"kubernetes.io/projected/a4c81032-5e90-4b7d-a27d-a46b7e052269-kube-api-access-t4vr6\") pod \"nova-metadata-0\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " pod="openstack/nova-metadata-0" Jan 26 09:26:10 crc kubenswrapper[4872]: I0126 09:26:10.892434 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:26:11 crc kubenswrapper[4872]: I0126 09:26:11.201627 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62d627e7-d04f-4673-a7c4-3d38488f0244" path="/var/lib/kubelet/pods/62d627e7-d04f-4673-a7c4-3d38488f0244/volumes" Jan 26 09:26:11 crc kubenswrapper[4872]: I0126 09:26:11.202866 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c09be6d1-5975-4b6c-9550-e4e329b6ae94" path="/var/lib/kubelet/pods/c09be6d1-5975-4b6c-9550-e4e329b6ae94/volumes" Jan 26 09:26:11 crc kubenswrapper[4872]: I0126 09:26:11.204589 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:26:11 crc kubenswrapper[4872]: W0126 09:26:11.209560 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19b7b80d_07f5_472b_b3df_dce1de4223a6.slice/crio-97e4826b92985e91c7d106f133a209b5d2aea078234c3f8911a9f1d281ff7093 WatchSource:0}: Error finding container 97e4826b92985e91c7d106f133a209b5d2aea078234c3f8911a9f1d281ff7093: Status 404 returned error can't find the container with id 97e4826b92985e91c7d106f133a209b5d2aea078234c3f8911a9f1d281ff7093 Jan 26 09:26:11 crc kubenswrapper[4872]: I0126 09:26:11.286279 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"dc5d286f-e785-4a4b-b96e-392ea79d7d2e","Type":"ContainerStarted","Data":"6cd26c08e6234085b5ab923349f1f6c954fca88c8c76cd07f7c0c6aac824be56"} Jan 26 09:26:11 crc kubenswrapper[4872]: I0126 09:26:11.286422 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"dc5d286f-e785-4a4b-b96e-392ea79d7d2e","Type":"ContainerStarted","Data":"ca0049935f85e434781633e8e54495946c182e25f9ccc62d6a62643bc15b5414"} Jan 26 09:26:11 crc kubenswrapper[4872]: I0126 09:26:11.286470 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:11 crc kubenswrapper[4872]: I0126 09:26:11.289731 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19b7b80d-07f5-472b-b3df-dce1de4223a6","Type":"ContainerStarted","Data":"97e4826b92985e91c7d106f133a209b5d2aea078234c3f8911a9f1d281ff7093"} Jan 26 09:26:11 crc kubenswrapper[4872]: I0126 09:26:11.304650 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.304625811 podStartE2EDuration="2.304625811s" podCreationTimestamp="2026-01-26 09:26:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:26:11.304157149 +0000 UTC m=+1104.612996950" watchObservedRunningTime="2026-01-26 09:26:11.304625811 +0000 UTC m=+1104.613465612" Jan 26 09:26:11 crc kubenswrapper[4872]: I0126 09:26:11.390687 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.011440 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.023246 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-combined-ca-bundle\") pod \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.023444 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-config-data\") pod \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.023551 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-logs\") pod \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.023732 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb8kb\" (UniqueName: \"kubernetes.io/projected/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-kube-api-access-sb8kb\") pod \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\" (UID: \"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6\") " Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.024036 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-logs" (OuterVolumeSpecName: "logs") pod "e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" (UID: "e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.024403 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.032068 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-kube-api-access-sb8kb" (OuterVolumeSpecName: "kube-api-access-sb8kb") pod "e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" (UID: "e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6"). InnerVolumeSpecName "kube-api-access-sb8kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.059938 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" (UID: "e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.084393 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-config-data" (OuterVolumeSpecName: "config-data") pod "e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" (UID: "e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.125619 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.125656 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb8kb\" (UniqueName: \"kubernetes.io/projected/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-kube-api-access-sb8kb\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.125671 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.305293 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19b7b80d-07f5-472b-b3df-dce1de4223a6","Type":"ContainerStarted","Data":"44e0355a8eff3442560d63c49f4c428c6bef0da33f9cfdb0fb7c6e586980d39c"} Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.308080 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a4c81032-5e90-4b7d-a27d-a46b7e052269","Type":"ContainerStarted","Data":"74d11d13b2a5c9ccf88d535de057db9ed169578d1016a5efdf2b773f425a746b"} Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.308119 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a4c81032-5e90-4b7d-a27d-a46b7e052269","Type":"ContainerStarted","Data":"569eedd9c241b70919ce97797c49d587e23e0812b018b6bb6b13d967dbfd30e3"} Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.308138 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a4c81032-5e90-4b7d-a27d-a46b7e052269","Type":"ContainerStarted","Data":"89c409f1221aa4dbfd7a0a9a598a09f35e2aa0e0baf96cb357e2210e17b3b558"} Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.310841 4872 generic.go:334] "Generic (PLEG): container finished" podID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" containerID="894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48" exitCode=0 Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.311858 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.317413 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6","Type":"ContainerDied","Data":"894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48"} Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.317484 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6","Type":"ContainerDied","Data":"7ccbdc1ff908d20b9b7b19f51265ca14f600c56528cd05e86526e1e9994952e5"} Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.317508 4872 scope.go:117] "RemoveContainer" containerID="894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.331088 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.331064025 podStartE2EDuration="2.331064025s" podCreationTimestamp="2026-01-26 09:26:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:26:12.320225347 +0000 UTC m=+1105.629065148" watchObservedRunningTime="2026-01-26 09:26:12.331064025 +0000 UTC m=+1105.639903826" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.367406 4872 scope.go:117] "RemoveContainer" containerID="ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.374923 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.37489023 podStartE2EDuration="2.37489023s" podCreationTimestamp="2026-01-26 09:26:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:26:12.356430776 +0000 UTC m=+1105.665270597" watchObservedRunningTime="2026-01-26 09:26:12.37489023 +0000 UTC m=+1105.683730041" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.405289 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.414054 4872 scope.go:117] "RemoveContainer" containerID="894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48" Jan 26 09:26:12 crc kubenswrapper[4872]: E0126 09:26:12.414704 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48\": container with ID starting with 894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48 not found: ID does not exist" containerID="894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.414743 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48"} err="failed to get container status \"894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48\": rpc error: code = NotFound desc = could not find container \"894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48\": container with ID starting with 894bb5d0a267a27edac37d6e58e95875c3b53008fcd5843c96415f081ef79e48 not found: ID does not exist" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.414771 4872 scope.go:117] "RemoveContainer" containerID="ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe" Jan 26 09:26:12 crc kubenswrapper[4872]: E0126 09:26:12.415303 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe\": container with ID starting with ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe not found: ID does not exist" containerID="ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.415338 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe"} err="failed to get container status \"ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe\": rpc error: code = NotFound desc = could not find container \"ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe\": container with ID starting with ceed6f05335a88665038a6c1372ce6b50099bc93c557862bfe8c01a5690f18fe not found: ID does not exist" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.418587 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.426635 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:12 crc kubenswrapper[4872]: E0126 09:26:12.427864 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" containerName="nova-api-api" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.427890 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" containerName="nova-api-api" Jan 26 09:26:12 crc kubenswrapper[4872]: E0126 09:26:12.427908 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" containerName="nova-api-log" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.427914 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" containerName="nova-api-log" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.428110 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" containerName="nova-api-log" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.428130 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" containerName="nova-api-api" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.434757 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.440346 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.466013 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.536917 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hn87j\" (UniqueName: \"kubernetes.io/projected/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-kube-api-access-hn87j\") pod \"nova-api-0\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.536998 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-config-data\") pod \"nova-api-0\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.537041 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-logs\") pod \"nova-api-0\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.537073 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.639303 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-config-data\") pod \"nova-api-0\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.639759 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-logs\") pod \"nova-api-0\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.639945 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.640170 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hn87j\" (UniqueName: \"kubernetes.io/projected/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-kube-api-access-hn87j\") pod \"nova-api-0\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.642049 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-logs\") pod \"nova-api-0\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.647175 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-config-data\") pod \"nova-api-0\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.648335 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.660689 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hn87j\" (UniqueName: \"kubernetes.io/projected/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-kube-api-access-hn87j\") pod \"nova-api-0\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " pod="openstack/nova-api-0" Jan 26 09:26:12 crc kubenswrapper[4872]: I0126 09:26:12.751464 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:26:13 crc kubenswrapper[4872]: I0126 09:26:13.197047 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6" path="/var/lib/kubelet/pods/e3dc6ee7-87c6-4e7e-b649-0006cb6feaf6/volumes" Jan 26 09:26:13 crc kubenswrapper[4872]: I0126 09:26:13.292061 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:13 crc kubenswrapper[4872]: W0126 09:26:13.296636 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode13988b3_053d_4eb0_9e8e_d87b32d4d9fe.slice/crio-c1da2bcb4bcada650f69e1dcf3b6fe87a3dad59e4731564cfb3b5f7a44aa1014 WatchSource:0}: Error finding container c1da2bcb4bcada650f69e1dcf3b6fe87a3dad59e4731564cfb3b5f7a44aa1014: Status 404 returned error can't find the container with id c1da2bcb4bcada650f69e1dcf3b6fe87a3dad59e4731564cfb3b5f7a44aa1014 Jan 26 09:26:13 crc kubenswrapper[4872]: I0126 09:26:13.324285 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe","Type":"ContainerStarted","Data":"c1da2bcb4bcada650f69e1dcf3b6fe87a3dad59e4731564cfb3b5f7a44aa1014"} Jan 26 09:26:14 crc kubenswrapper[4872]: I0126 09:26:14.340090 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe","Type":"ContainerStarted","Data":"a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a"} Jan 26 09:26:15 crc kubenswrapper[4872]: I0126 09:26:15.351251 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe","Type":"ContainerStarted","Data":"e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99"} Jan 26 09:26:15 crc kubenswrapper[4872]: I0126 09:26:15.383176 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.383152839 podStartE2EDuration="3.383152839s" podCreationTimestamp="2026-01-26 09:26:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:26:15.378173601 +0000 UTC m=+1108.687013412" watchObservedRunningTime="2026-01-26 09:26:15.383152839 +0000 UTC m=+1108.691992640" Jan 26 09:26:15 crc kubenswrapper[4872]: I0126 09:26:15.715944 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 26 09:26:15 crc kubenswrapper[4872]: I0126 09:26:15.893430 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 09:26:15 crc kubenswrapper[4872]: I0126 09:26:15.893494 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 09:26:20 crc kubenswrapper[4872]: I0126 09:26:20.034767 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Jan 26 09:26:20 crc kubenswrapper[4872]: I0126 09:26:20.716168 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 26 09:26:20 crc kubenswrapper[4872]: I0126 09:26:20.752755 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 26 09:26:20 crc kubenswrapper[4872]: I0126 09:26:20.893272 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 26 09:26:20 crc kubenswrapper[4872]: I0126 09:26:20.893323 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 26 09:26:21 crc kubenswrapper[4872]: I0126 09:26:21.454360 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 26 09:26:21 crc kubenswrapper[4872]: I0126 09:26:21.906000 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 09:26:21 crc kubenswrapper[4872]: I0126 09:26:21.906020 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 09:26:22 crc kubenswrapper[4872]: I0126 09:26:22.751971 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 09:26:22 crc kubenswrapper[4872]: I0126 09:26:22.752024 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 09:26:23 crc kubenswrapper[4872]: I0126 09:26:23.310962 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 26 09:26:23 crc kubenswrapper[4872]: I0126 09:26:23.439266 4872 generic.go:334] "Generic (PLEG): container finished" podID="e135958c-61c2-4199-94b3-2a8e6623310c" containerID="7d42cdc2ea0cb003d41f0df8602b37aab9808928402483a19fbde86b805272f0" exitCode=0 Jan 26 09:26:23 crc kubenswrapper[4872]: I0126 09:26:23.439320 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"e135958c-61c2-4199-94b3-2a8e6623310c","Type":"ContainerDied","Data":"7d42cdc2ea0cb003d41f0df8602b37aab9808928402483a19fbde86b805272f0"} Jan 26 09:26:23 crc kubenswrapper[4872]: I0126 09:26:23.834012 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.203:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 09:26:23 crc kubenswrapper[4872]: I0126 09:26:23.834087 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.203:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 09:26:24 crc kubenswrapper[4872]: I0126 09:26:24.455100 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"e135958c-61c2-4199-94b3-2a8e6623310c","Type":"ContainerStarted","Data":"312e0a6e68ff2038f0f9747616932ca1364df6a6d2a7b5a3ac8014f4e41fb913"} Jan 26 09:26:25 crc kubenswrapper[4872]: I0126 09:26:25.489962 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"e135958c-61c2-4199-94b3-2a8e6623310c","Type":"ContainerStarted","Data":"db67cddf43486c8a668d05a5af2103a4e6125c761e90be4905486e3a1702454c"} Jan 26 09:26:25 crc kubenswrapper[4872]: I0126 09:26:25.490680 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-conductor-0" Jan 26 09:26:25 crc kubenswrapper[4872]: I0126 09:26:25.490706 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"e135958c-61c2-4199-94b3-2a8e6623310c","Type":"ContainerStarted","Data":"1a1fa8308446733869fb4ee9be352ea86454b9a3325e5165c825209cfcf06e0a"} Jan 26 09:26:25 crc kubenswrapper[4872]: I0126 09:26:25.541013 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-conductor-0" podStartSLOduration=75.645219027 podStartE2EDuration="2m4.540985282s" podCreationTimestamp="2026-01-26 09:24:21 +0000 UTC" firstStartedPulling="2026-01-26 09:24:27.543514031 +0000 UTC m=+1000.852353832" lastFinishedPulling="2026-01-26 09:25:16.439280266 +0000 UTC m=+1049.748120087" observedRunningTime="2026-01-26 09:26:25.529568649 +0000 UTC m=+1118.838408460" watchObservedRunningTime="2026-01-26 09:26:25.540985282 +0000 UTC m=+1118.849825093" Jan 26 09:26:27 crc kubenswrapper[4872]: I0126 09:26:27.331210 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-conductor-0" Jan 26 09:26:27 crc kubenswrapper[4872]: I0126 09:26:27.576584 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 09:26:27 crc kubenswrapper[4872]: I0126 09:26:27.576900 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="3d16170b-8fc9-4c93-85fe-9ccef4da5bd0" containerName="kube-state-metrics" containerID="cri-o://bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696" gracePeriod=30 Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.181167 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.337088 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p98pt\" (UniqueName: \"kubernetes.io/projected/3d16170b-8fc9-4c93-85fe-9ccef4da5bd0-kube-api-access-p98pt\") pod \"3d16170b-8fc9-4c93-85fe-9ccef4da5bd0\" (UID: \"3d16170b-8fc9-4c93-85fe-9ccef4da5bd0\") " Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.344869 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d16170b-8fc9-4c93-85fe-9ccef4da5bd0-kube-api-access-p98pt" (OuterVolumeSpecName: "kube-api-access-p98pt") pod "3d16170b-8fc9-4c93-85fe-9ccef4da5bd0" (UID: "3d16170b-8fc9-4c93-85fe-9ccef4da5bd0"). InnerVolumeSpecName "kube-api-access-p98pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.439991 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p98pt\" (UniqueName: \"kubernetes.io/projected/3d16170b-8fc9-4c93-85fe-9ccef4da5bd0-kube-api-access-p98pt\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.523173 4872 generic.go:334] "Generic (PLEG): container finished" podID="3d16170b-8fc9-4c93-85fe-9ccef4da5bd0" containerID="bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696" exitCode=2 Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.523244 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.523730 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3d16170b-8fc9-4c93-85fe-9ccef4da5bd0","Type":"ContainerDied","Data":"bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696"} Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.523998 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3d16170b-8fc9-4c93-85fe-9ccef4da5bd0","Type":"ContainerDied","Data":"94e7d545c7f6e0ad17da0f7fe0de17d2c79b872267a79d0348cf739b5ce3ee7c"} Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.524089 4872 scope.go:117] "RemoveContainer" containerID="bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.570348 4872 scope.go:117] "RemoveContainer" containerID="bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696" Jan 26 09:26:28 crc kubenswrapper[4872]: E0126 09:26:28.574494 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696\": container with ID starting with bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696 not found: ID does not exist" containerID="bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.574556 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696"} err="failed to get container status \"bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696\": rpc error: code = NotFound desc = could not find container \"bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696\": container with ID starting with bca2ca5f2988a9e1cb2e055a70a7afc5f4cb67b35d7ab604394f4430e9cc8696 not found: ID does not exist" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.576550 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.598518 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.616576 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 09:26:28 crc kubenswrapper[4872]: E0126 09:26:28.617670 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d16170b-8fc9-4c93-85fe-9ccef4da5bd0" containerName="kube-state-metrics" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.617700 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d16170b-8fc9-4c93-85fe-9ccef4da5bd0" containerName="kube-state-metrics" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.617986 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d16170b-8fc9-4c93-85fe-9ccef4da5bd0" containerName="kube-state-metrics" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.618850 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.621397 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.621476 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.625260 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ironic-conductor-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.641050 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.746475 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2917f9e7-80a7-4af0-9816-aa58a9684f94-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2917f9e7-80a7-4af0-9816-aa58a9684f94\") " pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.746561 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2917f9e7-80a7-4af0-9816-aa58a9684f94-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2917f9e7-80a7-4af0-9816-aa58a9684f94\") " pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.746829 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2917f9e7-80a7-4af0-9816-aa58a9684f94-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2917f9e7-80a7-4af0-9816-aa58a9684f94\") " pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.747313 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffvc5\" (UniqueName: \"kubernetes.io/projected/2917f9e7-80a7-4af0-9816-aa58a9684f94-kube-api-access-ffvc5\") pod \"kube-state-metrics-0\" (UID: \"2917f9e7-80a7-4af0-9816-aa58a9684f94\") " pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.849514 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffvc5\" (UniqueName: \"kubernetes.io/projected/2917f9e7-80a7-4af0-9816-aa58a9684f94-kube-api-access-ffvc5\") pod \"kube-state-metrics-0\" (UID: \"2917f9e7-80a7-4af0-9816-aa58a9684f94\") " pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.849621 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2917f9e7-80a7-4af0-9816-aa58a9684f94-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2917f9e7-80a7-4af0-9816-aa58a9684f94\") " pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.849669 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2917f9e7-80a7-4af0-9816-aa58a9684f94-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2917f9e7-80a7-4af0-9816-aa58a9684f94\") " pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.849705 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2917f9e7-80a7-4af0-9816-aa58a9684f94-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2917f9e7-80a7-4af0-9816-aa58a9684f94\") " pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.854182 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2917f9e7-80a7-4af0-9816-aa58a9684f94-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2917f9e7-80a7-4af0-9816-aa58a9684f94\") " pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.856582 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2917f9e7-80a7-4af0-9816-aa58a9684f94-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2917f9e7-80a7-4af0-9816-aa58a9684f94\") " pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.860474 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2917f9e7-80a7-4af0-9816-aa58a9684f94-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2917f9e7-80a7-4af0-9816-aa58a9684f94\") " pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.869974 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffvc5\" (UniqueName: \"kubernetes.io/projected/2917f9e7-80a7-4af0-9816-aa58a9684f94-kube-api-access-ffvc5\") pod \"kube-state-metrics-0\" (UID: \"2917f9e7-80a7-4af0-9816-aa58a9684f94\") " pod="openstack/kube-state-metrics-0" Jan 26 09:26:28 crc kubenswrapper[4872]: I0126 09:26:28.956125 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Jan 26 09:26:29 crc kubenswrapper[4872]: I0126 09:26:29.207917 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d16170b-8fc9-4c93-85fe-9ccef4da5bd0" path="/var/lib/kubelet/pods/3d16170b-8fc9-4c93-85fe-9ccef4da5bd0/volumes" Jan 26 09:26:29 crc kubenswrapper[4872]: W0126 09:26:29.551597 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2917f9e7_80a7_4af0_9816_aa58a9684f94.slice/crio-8240eb47c0abda6cc2a9c38b471176567dd57f4c0e7d0b9edda8ba1ab668c4fb WatchSource:0}: Error finding container 8240eb47c0abda6cc2a9c38b471176567dd57f4c0e7d0b9edda8ba1ab668c4fb: Status 404 returned error can't find the container with id 8240eb47c0abda6cc2a9c38b471176567dd57f4c0e7d0b9edda8ba1ab668c4fb Jan 26 09:26:29 crc kubenswrapper[4872]: I0126 09:26:29.555182 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Jan 26 09:26:29 crc kubenswrapper[4872]: I0126 09:26:29.607478 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:26:29 crc kubenswrapper[4872]: I0126 09:26:29.607854 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="ceilometer-central-agent" containerID="cri-o://ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a" gracePeriod=30 Jan 26 09:26:29 crc kubenswrapper[4872]: I0126 09:26:29.607986 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="sg-core" containerID="cri-o://175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174" gracePeriod=30 Jan 26 09:26:29 crc kubenswrapper[4872]: I0126 09:26:29.608015 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="ceilometer-notification-agent" containerID="cri-o://8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6" gracePeriod=30 Jan 26 09:26:29 crc kubenswrapper[4872]: I0126 09:26:29.608235 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="proxy-httpd" containerID="cri-o://0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4" gracePeriod=30 Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.549928 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2917f9e7-80a7-4af0-9816-aa58a9684f94","Type":"ContainerStarted","Data":"55a66af03cdc57ca67814a4361fe1c0f7db3a1bade914b47dd2fdca720428e69"} Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.550954 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2917f9e7-80a7-4af0-9816-aa58a9684f94","Type":"ContainerStarted","Data":"8240eb47c0abda6cc2a9c38b471176567dd57f4c0e7d0b9edda8ba1ab668c4fb"} Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.552508 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.557551 4872 generic.go:334] "Generic (PLEG): container finished" podID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerID="0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4" exitCode=0 Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.557585 4872 generic.go:334] "Generic (PLEG): container finished" podID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerID="175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174" exitCode=2 Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.557592 4872 generic.go:334] "Generic (PLEG): container finished" podID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerID="ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a" exitCode=0 Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.557617 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b378cb-a416-449b-8c1e-0a052aba75bd","Type":"ContainerDied","Data":"0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4"} Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.557645 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b378cb-a416-449b-8c1e-0a052aba75bd","Type":"ContainerDied","Data":"175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174"} Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.557654 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b378cb-a416-449b-8c1e-0a052aba75bd","Type":"ContainerDied","Data":"ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a"} Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.576025 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.213787585 podStartE2EDuration="2.576003526s" podCreationTimestamp="2026-01-26 09:26:28 +0000 UTC" firstStartedPulling="2026-01-26 09:26:29.554770106 +0000 UTC m=+1122.863609907" lastFinishedPulling="2026-01-26 09:26:29.916986047 +0000 UTC m=+1123.225825848" observedRunningTime="2026-01-26 09:26:30.568383841 +0000 UTC m=+1123.877223652" watchObservedRunningTime="2026-01-26 09:26:30.576003526 +0000 UTC m=+1123.884843327" Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.898788 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.903279 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 26 09:26:30 crc kubenswrapper[4872]: I0126 09:26:30.919362 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 26 09:26:31 crc kubenswrapper[4872]: I0126 09:26:31.580357 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.573773 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.584910 4872 generic.go:334] "Generic (PLEG): container finished" podID="14a273ac-90fd-42dd-b5f0-1fd7ca8283fe" containerID="898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb" exitCode=137 Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.584959 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe","Type":"ContainerDied","Data":"898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb"} Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.584991 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.585007 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe","Type":"ContainerDied","Data":"39c4eb0be1328947c7748abf61ec7eb6aa6b748595ea9d37155a2a0c7308ad36"} Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.585040 4872 scope.go:117] "RemoveContainer" containerID="898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.619515 4872 scope.go:117] "RemoveContainer" containerID="898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb" Jan 26 09:26:32 crc kubenswrapper[4872]: E0126 09:26:32.620140 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb\": container with ID starting with 898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb not found: ID does not exist" containerID="898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.620168 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb"} err="failed to get container status \"898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb\": rpc error: code = NotFound desc = could not find container \"898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb\": container with ID starting with 898970c07fe6635d942c36b3f59dcd7f9baa192132ac213221c988cbc98e97cb not found: ID does not exist" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.643761 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w6d7\" (UniqueName: \"kubernetes.io/projected/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-kube-api-access-9w6d7\") pod \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\" (UID: \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\") " Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.643981 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-config-data\") pod \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\" (UID: \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\") " Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.644028 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-combined-ca-bundle\") pod \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\" (UID: \"14a273ac-90fd-42dd-b5f0-1fd7ca8283fe\") " Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.651069 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-kube-api-access-9w6d7" (OuterVolumeSpecName: "kube-api-access-9w6d7") pod "14a273ac-90fd-42dd-b5f0-1fd7ca8283fe" (UID: "14a273ac-90fd-42dd-b5f0-1fd7ca8283fe"). InnerVolumeSpecName "kube-api-access-9w6d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.675390 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14a273ac-90fd-42dd-b5f0-1fd7ca8283fe" (UID: "14a273ac-90fd-42dd-b5f0-1fd7ca8283fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.680852 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-config-data" (OuterVolumeSpecName: "config-data") pod "14a273ac-90fd-42dd-b5f0-1fd7ca8283fe" (UID: "14a273ac-90fd-42dd-b5f0-1fd7ca8283fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.746759 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.746825 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.746845 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w6d7\" (UniqueName: \"kubernetes.io/projected/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe-kube-api-access-9w6d7\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.756302 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.757142 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.759882 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.763426 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.920542 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.935327 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.953326 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 09:26:32 crc kubenswrapper[4872]: E0126 09:26:32.953977 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14a273ac-90fd-42dd-b5f0-1fd7ca8283fe" containerName="nova-cell1-novncproxy-novncproxy" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.954011 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="14a273ac-90fd-42dd-b5f0-1fd7ca8283fe" containerName="nova-cell1-novncproxy-novncproxy" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.954274 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="14a273ac-90fd-42dd-b5f0-1fd7ca8283fe" containerName="nova-cell1-novncproxy-novncproxy" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.955182 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.963607 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.966169 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.966382 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Jan 26 09:26:32 crc kubenswrapper[4872]: I0126 09:26:32.967605 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.050698 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8903a57e-a76b-42a5-a025-992c12eab47c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.050745 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8903a57e-a76b-42a5-a025-992c12eab47c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.050822 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjkm5\" (UniqueName: \"kubernetes.io/projected/8903a57e-a76b-42a5-a025-992c12eab47c-kube-api-access-bjkm5\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.050889 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8903a57e-a76b-42a5-a025-992c12eab47c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.050931 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8903a57e-a76b-42a5-a025-992c12eab47c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.153941 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8903a57e-a76b-42a5-a025-992c12eab47c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.154415 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8903a57e-a76b-42a5-a025-992c12eab47c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.154517 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjkm5\" (UniqueName: \"kubernetes.io/projected/8903a57e-a76b-42a5-a025-992c12eab47c-kube-api-access-bjkm5\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.154560 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8903a57e-a76b-42a5-a025-992c12eab47c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.154617 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8903a57e-a76b-42a5-a025-992c12eab47c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.159678 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8903a57e-a76b-42a5-a025-992c12eab47c-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.160898 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8903a57e-a76b-42a5-a025-992c12eab47c-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.161299 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8903a57e-a76b-42a5-a025-992c12eab47c-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.161572 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8903a57e-a76b-42a5-a025-992c12eab47c-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.177445 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjkm5\" (UniqueName: \"kubernetes.io/projected/8903a57e-a76b-42a5-a025-992c12eab47c-kube-api-access-bjkm5\") pod \"nova-cell1-novncproxy-0\" (UID: \"8903a57e-a76b-42a5-a025-992c12eab47c\") " pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.199266 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14a273ac-90fd-42dd-b5f0-1fd7ca8283fe" path="/var/lib/kubelet/pods/14a273ac-90fd-42dd-b5f0-1fd7ca8283fe/volumes" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.282237 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.592962 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.600321 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.610973 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.783459 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-nzwnv"] Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.785278 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.793340 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-nzwnv"] Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.892996 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.893080 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.893103 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.893219 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq6p9\" (UniqueName: \"kubernetes.io/projected/4654462c-cb3b-46d4-aec6-9e701038aebd-kube-api-access-lq6p9\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.893270 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-config\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.893529 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.995518 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.995606 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.995627 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.995657 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq6p9\" (UniqueName: \"kubernetes.io/projected/4654462c-cb3b-46d4-aec6-9e701038aebd-kube-api-access-lq6p9\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.995683 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-config\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.995767 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.996834 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.996860 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.996949 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.997377 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-config\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:33 crc kubenswrapper[4872]: I0126 09:26:33.997740 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4654462c-cb3b-46d4-aec6-9e701038aebd-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.022256 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq6p9\" (UniqueName: \"kubernetes.io/projected/4654462c-cb3b-46d4-aec6-9e701038aebd-kube-api-access-lq6p9\") pod \"dnsmasq-dns-59cf4bdb65-nzwnv\" (UID: \"4654462c-cb3b-46d4-aec6-9e701038aebd\") " pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.130845 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.341049 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.402994 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-combined-ca-bundle\") pod \"65b378cb-a416-449b-8c1e-0a052aba75bd\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.403147 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt975\" (UniqueName: \"kubernetes.io/projected/65b378cb-a416-449b-8c1e-0a052aba75bd-kube-api-access-tt975\") pod \"65b378cb-a416-449b-8c1e-0a052aba75bd\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.403272 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b378cb-a416-449b-8c1e-0a052aba75bd-run-httpd\") pod \"65b378cb-a416-449b-8c1e-0a052aba75bd\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.403376 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-config-data\") pod \"65b378cb-a416-449b-8c1e-0a052aba75bd\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.403413 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b378cb-a416-449b-8c1e-0a052aba75bd-log-httpd\") pod \"65b378cb-a416-449b-8c1e-0a052aba75bd\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.403452 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-sg-core-conf-yaml\") pod \"65b378cb-a416-449b-8c1e-0a052aba75bd\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.403507 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-scripts\") pod \"65b378cb-a416-449b-8c1e-0a052aba75bd\" (UID: \"65b378cb-a416-449b-8c1e-0a052aba75bd\") " Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.406293 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65b378cb-a416-449b-8c1e-0a052aba75bd-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "65b378cb-a416-449b-8c1e-0a052aba75bd" (UID: "65b378cb-a416-449b-8c1e-0a052aba75bd"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.406974 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65b378cb-a416-449b-8c1e-0a052aba75bd-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "65b378cb-a416-449b-8c1e-0a052aba75bd" (UID: "65b378cb-a416-449b-8c1e-0a052aba75bd"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.444711 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-scripts" (OuterVolumeSpecName: "scripts") pod "65b378cb-a416-449b-8c1e-0a052aba75bd" (UID: "65b378cb-a416-449b-8c1e-0a052aba75bd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.445050 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65b378cb-a416-449b-8c1e-0a052aba75bd-kube-api-access-tt975" (OuterVolumeSpecName: "kube-api-access-tt975") pod "65b378cb-a416-449b-8c1e-0a052aba75bd" (UID: "65b378cb-a416-449b-8c1e-0a052aba75bd"). InnerVolumeSpecName "kube-api-access-tt975". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.456996 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "65b378cb-a416-449b-8c1e-0a052aba75bd" (UID: "65b378cb-a416-449b-8c1e-0a052aba75bd"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.518269 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b378cb-a416-449b-8c1e-0a052aba75bd-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.518301 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.521026 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.521059 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt975\" (UniqueName: \"kubernetes.io/projected/65b378cb-a416-449b-8c1e-0a052aba75bd-kube-api-access-tt975\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.521070 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/65b378cb-a416-449b-8c1e-0a052aba75bd-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.551724 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65b378cb-a416-449b-8c1e-0a052aba75bd" (UID: "65b378cb-a416-449b-8c1e-0a052aba75bd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.591116 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-config-data" (OuterVolumeSpecName: "config-data") pod "65b378cb-a416-449b-8c1e-0a052aba75bd" (UID: "65b378cb-a416-449b-8c1e-0a052aba75bd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.616182 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8903a57e-a76b-42a5-a025-992c12eab47c","Type":"ContainerStarted","Data":"54ac6052b6b3b50ba55bbe4696c56833b343ea1eb9b8e40dc59f0771c3f80f8e"} Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.616235 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8903a57e-a76b-42a5-a025-992c12eab47c","Type":"ContainerStarted","Data":"81d680367b8d3b8becfca8b6225570a22b6387b7aadda0046ed76977d66a786f"} Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.621345 4872 generic.go:334] "Generic (PLEG): container finished" podID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerID="8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6" exitCode=0 Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.621985 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.622316 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b378cb-a416-449b-8c1e-0a052aba75bd","Type":"ContainerDied","Data":"8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6"} Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.622338 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"65b378cb-a416-449b-8c1e-0a052aba75bd","Type":"ContainerDied","Data":"024400034c51e341443484d2b17631ce989e5f9f7c68f352ac1e6bc2c3c1d7c6"} Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.622355 4872 scope.go:117] "RemoveContainer" containerID="0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.626426 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.626540 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65b378cb-a416-449b-8c1e-0a052aba75bd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.639272 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.639246331 podStartE2EDuration="2.639246331s" podCreationTimestamp="2026-01-26 09:26:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:26:34.633459012 +0000 UTC m=+1127.942298813" watchObservedRunningTime="2026-01-26 09:26:34.639246331 +0000 UTC m=+1127.948086132" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.647012 4872 scope.go:117] "RemoveContainer" containerID="175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.681429 4872 scope.go:117] "RemoveContainer" containerID="8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.714422 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.720967 4872 scope.go:117] "RemoveContainer" containerID="ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.738283 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.753117 4872 scope.go:117] "RemoveContainer" containerID="0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4" Jan 26 09:26:34 crc kubenswrapper[4872]: E0126 09:26:34.753649 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4\": container with ID starting with 0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4 not found: ID does not exist" containerID="0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.753684 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4"} err="failed to get container status \"0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4\": rpc error: code = NotFound desc = could not find container \"0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4\": container with ID starting with 0b477500d336667fe1f647256a49cd45ffca14b80fbf239fe576071077dd8de4 not found: ID does not exist" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.753728 4872 scope.go:117] "RemoveContainer" containerID="175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174" Jan 26 09:26:34 crc kubenswrapper[4872]: E0126 09:26:34.753916 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174\": container with ID starting with 175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174 not found: ID does not exist" containerID="175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.753933 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174"} err="failed to get container status \"175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174\": rpc error: code = NotFound desc = could not find container \"175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174\": container with ID starting with 175b2f5ea74ba703f2b47052915bf93c5764b685bdc7b6be74f04a7a8f528174 not found: ID does not exist" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.753966 4872 scope.go:117] "RemoveContainer" containerID="8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6" Jan 26 09:26:34 crc kubenswrapper[4872]: E0126 09:26:34.754152 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6\": container with ID starting with 8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6 not found: ID does not exist" containerID="8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.754173 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6"} err="failed to get container status \"8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6\": rpc error: code = NotFound desc = could not find container \"8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6\": container with ID starting with 8e8fcc8b4288659e299a3b899e9e162e0502c247c4cb4f45221c27d8bd0381e6 not found: ID does not exist" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.754209 4872 scope.go:117] "RemoveContainer" containerID="ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a" Jan 26 09:26:34 crc kubenswrapper[4872]: E0126 09:26:34.754402 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a\": container with ID starting with ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a not found: ID does not exist" containerID="ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.754418 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a"} err="failed to get container status \"ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a\": rpc error: code = NotFound desc = could not find container \"ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a\": container with ID starting with ca2b0db8cb680ea13f25882f5730d271b60aa7ecc640c064f221b69e2eb6b68a not found: ID does not exist" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.755081 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:26:34 crc kubenswrapper[4872]: E0126 09:26:34.755472 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="proxy-httpd" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.755489 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="proxy-httpd" Jan 26 09:26:34 crc kubenswrapper[4872]: E0126 09:26:34.755507 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="ceilometer-notification-agent" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.755540 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="ceilometer-notification-agent" Jan 26 09:26:34 crc kubenswrapper[4872]: E0126 09:26:34.755559 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="ceilometer-central-agent" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.755565 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="ceilometer-central-agent" Jan 26 09:26:34 crc kubenswrapper[4872]: E0126 09:26:34.755571 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="sg-core" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.755577 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="sg-core" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.755765 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="proxy-httpd" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.755776 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="ceilometer-central-agent" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.755789 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="sg-core" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.755835 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" containerName="ceilometer-notification-agent" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.757741 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.763407 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.763888 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.764086 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.787220 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.805200 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-nzwnv"] Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.832104 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-config-data\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.832165 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed8f02d7-cd7c-4b04-9760-fb107c585017-log-httpd\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.832246 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.832271 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.832438 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5p47\" (UniqueName: \"kubernetes.io/projected/ed8f02d7-cd7c-4b04-9760-fb107c585017-kube-api-access-g5p47\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.832611 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.832674 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-scripts\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.832726 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed8f02d7-cd7c-4b04-9760-fb107c585017-run-httpd\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.935239 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-config-data\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.936105 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed8f02d7-cd7c-4b04-9760-fb107c585017-log-httpd\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.936159 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed8f02d7-cd7c-4b04-9760-fb107c585017-log-httpd\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.936237 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.936267 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.936670 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5p47\" (UniqueName: \"kubernetes.io/projected/ed8f02d7-cd7c-4b04-9760-fb107c585017-kube-api-access-g5p47\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.936755 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.936785 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-scripts\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.936839 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed8f02d7-cd7c-4b04-9760-fb107c585017-run-httpd\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.937298 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed8f02d7-cd7c-4b04-9760-fb107c585017-run-httpd\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.942444 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.943267 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.944009 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.944067 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-config-data\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.944836 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-scripts\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:34 crc kubenswrapper[4872]: I0126 09:26:34.957238 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5p47\" (UniqueName: \"kubernetes.io/projected/ed8f02d7-cd7c-4b04-9760-fb107c585017-kube-api-access-g5p47\") pod \"ceilometer-0\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " pod="openstack/ceilometer-0" Jan 26 09:26:35 crc kubenswrapper[4872]: I0126 09:26:35.087877 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:26:35 crc kubenswrapper[4872]: I0126 09:26:35.223353 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65b378cb-a416-449b-8c1e-0a052aba75bd" path="/var/lib/kubelet/pods/65b378cb-a416-449b-8c1e-0a052aba75bd/volumes" Jan 26 09:26:35 crc kubenswrapper[4872]: I0126 09:26:35.542914 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:26:35 crc kubenswrapper[4872]: I0126 09:26:35.617142 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:26:35 crc kubenswrapper[4872]: I0126 09:26:35.631963 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 09:26:35 crc kubenswrapper[4872]: I0126 09:26:35.639553 4872 generic.go:334] "Generic (PLEG): container finished" podID="4654462c-cb3b-46d4-aec6-9e701038aebd" containerID="845ebfe53a7153766c8f191f61cf5cdb2cbf71e0f92e98769ab6a1f66ad31b20" exitCode=0 Jan 26 09:26:35 crc kubenswrapper[4872]: I0126 09:26:35.639623 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" event={"ID":"4654462c-cb3b-46d4-aec6-9e701038aebd","Type":"ContainerDied","Data":"845ebfe53a7153766c8f191f61cf5cdb2cbf71e0f92e98769ab6a1f66ad31b20"} Jan 26 09:26:35 crc kubenswrapper[4872]: I0126 09:26:35.639658 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" event={"ID":"4654462c-cb3b-46d4-aec6-9e701038aebd","Type":"ContainerStarted","Data":"cf7e30c758bd3a23052ed9e84b9162895a2bae575535d22f4fc7a435892d60e8"} Jan 26 09:26:36 crc kubenswrapper[4872]: I0126 09:26:36.521010 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:36 crc kubenswrapper[4872]: I0126 09:26:36.653952 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" event={"ID":"4654462c-cb3b-46d4-aec6-9e701038aebd","Type":"ContainerStarted","Data":"9941b591d977f920ffd45cb3cf48730172465ec5c0695b3742f041ebcd8afd48"} Jan 26 09:26:36 crc kubenswrapper[4872]: I0126 09:26:36.654110 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:36 crc kubenswrapper[4872]: I0126 09:26:36.663963 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed8f02d7-cd7c-4b04-9760-fb107c585017","Type":"ContainerStarted","Data":"ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b"} Jan 26 09:26:36 crc kubenswrapper[4872]: I0126 09:26:36.664032 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed8f02d7-cd7c-4b04-9760-fb107c585017","Type":"ContainerStarted","Data":"8bd7e6fac1d13bf2be8686e2474a14ca44e69e8421d6d1046c07a198bcb0ffe6"} Jan 26 09:26:36 crc kubenswrapper[4872]: I0126 09:26:36.664125 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" containerName="nova-api-log" containerID="cri-o://a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a" gracePeriod=30 Jan 26 09:26:36 crc kubenswrapper[4872]: I0126 09:26:36.664230 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" containerName="nova-api-api" containerID="cri-o://e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99" gracePeriod=30 Jan 26 09:26:36 crc kubenswrapper[4872]: I0126 09:26:36.687314 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" podStartSLOduration=3.687284444 podStartE2EDuration="3.687284444s" podCreationTimestamp="2026-01-26 09:26:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:26:36.673963623 +0000 UTC m=+1129.982803414" watchObservedRunningTime="2026-01-26 09:26:36.687284444 +0000 UTC m=+1129.996124245" Jan 26 09:26:37 crc kubenswrapper[4872]: I0126 09:26:37.685836 4872 generic.go:334] "Generic (PLEG): container finished" podID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" containerID="a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a" exitCode=143 Jan 26 09:26:37 crc kubenswrapper[4872]: I0126 09:26:37.689059 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe","Type":"ContainerDied","Data":"a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a"} Jan 26 09:26:37 crc kubenswrapper[4872]: I0126 09:26:37.701275 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed8f02d7-cd7c-4b04-9760-fb107c585017","Type":"ContainerStarted","Data":"65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44"} Jan 26 09:26:38 crc kubenswrapper[4872]: I0126 09:26:38.283165 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:38 crc kubenswrapper[4872]: I0126 09:26:38.713673 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed8f02d7-cd7c-4b04-9760-fb107c585017","Type":"ContainerStarted","Data":"8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9"} Jan 26 09:26:38 crc kubenswrapper[4872]: I0126 09:26:38.971214 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Jan 26 09:26:39 crc kubenswrapper[4872]: I0126 09:26:39.734871 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed8f02d7-cd7c-4b04-9760-fb107c585017","Type":"ContainerStarted","Data":"5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4"} Jan 26 09:26:39 crc kubenswrapper[4872]: I0126 09:26:39.737317 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 09:26:39 crc kubenswrapper[4872]: I0126 09:26:39.736048 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="sg-core" containerID="cri-o://8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9" gracePeriod=30 Jan 26 09:26:39 crc kubenswrapper[4872]: I0126 09:26:39.736127 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="proxy-httpd" containerID="cri-o://5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4" gracePeriod=30 Jan 26 09:26:39 crc kubenswrapper[4872]: I0126 09:26:39.736189 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="ceilometer-notification-agent" containerID="cri-o://65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44" gracePeriod=30 Jan 26 09:26:39 crc kubenswrapper[4872]: I0126 09:26:39.735945 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="ceilometer-central-agent" containerID="cri-o://ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b" gracePeriod=30 Jan 26 09:26:39 crc kubenswrapper[4872]: I0126 09:26:39.780554 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.325226358 podStartE2EDuration="5.780523453s" podCreationTimestamp="2026-01-26 09:26:34 +0000 UTC" firstStartedPulling="2026-01-26 09:26:35.631576479 +0000 UTC m=+1128.940416280" lastFinishedPulling="2026-01-26 09:26:39.086873554 +0000 UTC m=+1132.395713375" observedRunningTime="2026-01-26 09:26:39.771498502 +0000 UTC m=+1133.080338383" watchObservedRunningTime="2026-01-26 09:26:39.780523453 +0000 UTC m=+1133.089363264" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.227827 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.228734 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.327761 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.376347 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-logs\") pod \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.376440 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-config-data\") pod \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.376459 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-combined-ca-bundle\") pod \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.376520 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hn87j\" (UniqueName: \"kubernetes.io/projected/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-kube-api-access-hn87j\") pod \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\" (UID: \"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe\") " Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.377055 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-logs" (OuterVolumeSpecName: "logs") pod "e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" (UID: "e13988b3-053d-4eb0-9e8e-d87b32d4d9fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.377279 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.388235 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-kube-api-access-hn87j" (OuterVolumeSpecName: "kube-api-access-hn87j") pod "e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" (UID: "e13988b3-053d-4eb0-9e8e-d87b32d4d9fe"). InnerVolumeSpecName "kube-api-access-hn87j". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.453662 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-config-data" (OuterVolumeSpecName: "config-data") pod "e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" (UID: "e13988b3-053d-4eb0-9e8e-d87b32d4d9fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.465610 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" (UID: "e13988b3-053d-4eb0-9e8e-d87b32d4d9fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.479092 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.479120 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.479134 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hn87j\" (UniqueName: \"kubernetes.io/projected/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe-kube-api-access-hn87j\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.749284 4872 generic.go:334] "Generic (PLEG): container finished" podID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" containerID="e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99" exitCode=0 Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.749443 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.749479 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe","Type":"ContainerDied","Data":"e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99"} Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.750395 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e13988b3-053d-4eb0-9e8e-d87b32d4d9fe","Type":"ContainerDied","Data":"c1da2bcb4bcada650f69e1dcf3b6fe87a3dad59e4731564cfb3b5f7a44aa1014"} Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.750450 4872 scope.go:117] "RemoveContainer" containerID="e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.763261 4872 generic.go:334] "Generic (PLEG): container finished" podID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerID="5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4" exitCode=0 Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.763298 4872 generic.go:334] "Generic (PLEG): container finished" podID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerID="8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9" exitCode=2 Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.763306 4872 generic.go:334] "Generic (PLEG): container finished" podID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerID="65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44" exitCode=0 Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.763330 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed8f02d7-cd7c-4b04-9760-fb107c585017","Type":"ContainerDied","Data":"5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4"} Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.763364 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed8f02d7-cd7c-4b04-9760-fb107c585017","Type":"ContainerDied","Data":"8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9"} Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.763374 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed8f02d7-cd7c-4b04-9760-fb107c585017","Type":"ContainerDied","Data":"65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44"} Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.788143 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.788340 4872 scope.go:117] "RemoveContainer" containerID="a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.822674 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.847050 4872 scope.go:117] "RemoveContainer" containerID="e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99" Jan 26 09:26:40 crc kubenswrapper[4872]: E0126 09:26:40.847940 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99\": container with ID starting with e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99 not found: ID does not exist" containerID="e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.847990 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99"} err="failed to get container status \"e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99\": rpc error: code = NotFound desc = could not find container \"e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99\": container with ID starting with e4a7545900fc581b39ab99cd65327d1f86386ad038be6d8a44bb2bd481287d99 not found: ID does not exist" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.848029 4872 scope.go:117] "RemoveContainer" containerID="a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a" Jan 26 09:26:40 crc kubenswrapper[4872]: E0126 09:26:40.848507 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a\": container with ID starting with a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a not found: ID does not exist" containerID="a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.848558 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a"} err="failed to get container status \"a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a\": rpc error: code = NotFound desc = could not find container \"a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a\": container with ID starting with a5c33d50d01ad3f4b013835e08beb51034999dabeeb1cf3a82acd42ce105983a not found: ID does not exist" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.864168 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:40 crc kubenswrapper[4872]: E0126 09:26:40.865592 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" containerName="nova-api-log" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.865617 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" containerName="nova-api-log" Jan 26 09:26:40 crc kubenswrapper[4872]: E0126 09:26:40.865649 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" containerName="nova-api-api" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.865657 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" containerName="nova-api-api" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.866051 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" containerName="nova-api-log" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.866119 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" containerName="nova-api-api" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.868183 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.877831 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.878093 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.878210 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.881548 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.991256 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-logs\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.991339 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6s762\" (UniqueName: \"kubernetes.io/projected/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-kube-api-access-6s762\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.991368 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-config-data\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.991427 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.991467 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-public-tls-certs\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:40 crc kubenswrapper[4872]: I0126 09:26:40.992133 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.094406 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.094470 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-logs\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.094500 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6s762\" (UniqueName: \"kubernetes.io/projected/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-kube-api-access-6s762\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.094518 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-config-data\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.094552 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.094582 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-public-tls-certs\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.098500 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-logs\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.101135 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-public-tls-certs\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.101643 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.102307 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-config-data\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.102611 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.119303 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6s762\" (UniqueName: \"kubernetes.io/projected/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-kube-api-access-6s762\") pod \"nova-api-0\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.193570 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.219967 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e13988b3-053d-4eb0-9e8e-d87b32d4d9fe" path="/var/lib/kubelet/pods/e13988b3-053d-4eb0-9e8e-d87b32d4d9fe/volumes" Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.750885 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:41 crc kubenswrapper[4872]: W0126 09:26:41.758741 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf728fc3d_3a48_4c25_b2ac_6d1d6ef0aa55.slice/crio-e9062b2d32ee86f79034f94f8bbfb74c07bc83b645c6b099a962ffa6b0ce5787 WatchSource:0}: Error finding container e9062b2d32ee86f79034f94f8bbfb74c07bc83b645c6b099a962ffa6b0ce5787: Status 404 returned error can't find the container with id e9062b2d32ee86f79034f94f8bbfb74c07bc83b645c6b099a962ffa6b0ce5787 Jan 26 09:26:41 crc kubenswrapper[4872]: I0126 09:26:41.779153 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55","Type":"ContainerStarted","Data":"e9062b2d32ee86f79034f94f8bbfb74c07bc83b645c6b099a962ffa6b0ce5787"} Jan 26 09:26:42 crc kubenswrapper[4872]: I0126 09:26:42.800735 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55","Type":"ContainerStarted","Data":"44b717b6cbdce542f10ea68ad34a469f7fea7c2ac580e2d6797e0d34a3464b6a"} Jan 26 09:26:42 crc kubenswrapper[4872]: I0126 09:26:42.801740 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55","Type":"ContainerStarted","Data":"52eb3db4a582c6797b9128c5130fd02f64f645682d724efd67e13f4932758626"} Jan 26 09:26:42 crc kubenswrapper[4872]: I0126 09:26:42.846993 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.846959545 podStartE2EDuration="2.846959545s" podCreationTimestamp="2026-01-26 09:26:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:26:42.826782397 +0000 UTC m=+1136.135622268" watchObservedRunningTime="2026-01-26 09:26:42.846959545 +0000 UTC m=+1136.155799386" Jan 26 09:26:43 crc kubenswrapper[4872]: I0126 09:26:43.282750 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:43 crc kubenswrapper[4872]: I0126 09:26:43.312983 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:43 crc kubenswrapper[4872]: I0126 09:26:43.838687 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.053433 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-rmm5v"] Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.055406 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.059255 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.059389 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.067845 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-scripts\") pod \"nova-cell1-cell-mapping-rmm5v\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.068033 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb5hw\" (UniqueName: \"kubernetes.io/projected/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-kube-api-access-hb5hw\") pod \"nova-cell1-cell-mapping-rmm5v\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.068079 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-config-data\") pod \"nova-cell1-cell-mapping-rmm5v\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.068120 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rmm5v\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.085896 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-rmm5v"] Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.132980 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-nzwnv" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.174344 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-scripts\") pod \"nova-cell1-cell-mapping-rmm5v\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.174666 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb5hw\" (UniqueName: \"kubernetes.io/projected/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-kube-api-access-hb5hw\") pod \"nova-cell1-cell-mapping-rmm5v\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.174707 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-config-data\") pod \"nova-cell1-cell-mapping-rmm5v\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.174789 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rmm5v\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.209498 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-scripts\") pod \"nova-cell1-cell-mapping-rmm5v\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.210474 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-gthgc"] Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.211044 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" podUID="0855d9b8-26d6-4175-beed-a85fecaa3d29" containerName="dnsmasq-dns" containerID="cri-o://ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6" gracePeriod=10 Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.211942 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb5hw\" (UniqueName: \"kubernetes.io/projected/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-kube-api-access-hb5hw\") pod \"nova-cell1-cell-mapping-rmm5v\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.215270 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-config-data\") pod \"nova-cell1-cell-mapping-rmm5v\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.215703 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-rmm5v\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.456437 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.468362 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.483700 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed8f02d7-cd7c-4b04-9760-fb107c585017-run-httpd\") pod \"ed8f02d7-cd7c-4b04-9760-fb107c585017\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.483791 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-config-data\") pod \"ed8f02d7-cd7c-4b04-9760-fb107c585017\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.483877 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-ceilometer-tls-certs\") pod \"ed8f02d7-cd7c-4b04-9760-fb107c585017\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.483908 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-sg-core-conf-yaml\") pod \"ed8f02d7-cd7c-4b04-9760-fb107c585017\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.483945 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-scripts\") pod \"ed8f02d7-cd7c-4b04-9760-fb107c585017\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.483993 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5p47\" (UniqueName: \"kubernetes.io/projected/ed8f02d7-cd7c-4b04-9760-fb107c585017-kube-api-access-g5p47\") pod \"ed8f02d7-cd7c-4b04-9760-fb107c585017\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.484061 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed8f02d7-cd7c-4b04-9760-fb107c585017-log-httpd\") pod \"ed8f02d7-cd7c-4b04-9760-fb107c585017\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.484149 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-combined-ca-bundle\") pod \"ed8f02d7-cd7c-4b04-9760-fb107c585017\" (UID: \"ed8f02d7-cd7c-4b04-9760-fb107c585017\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.485383 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed8f02d7-cd7c-4b04-9760-fb107c585017-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ed8f02d7-cd7c-4b04-9760-fb107c585017" (UID: "ed8f02d7-cd7c-4b04-9760-fb107c585017"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.485905 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed8f02d7-cd7c-4b04-9760-fb107c585017-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ed8f02d7-cd7c-4b04-9760-fb107c585017" (UID: "ed8f02d7-cd7c-4b04-9760-fb107c585017"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.493309 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed8f02d7-cd7c-4b04-9760-fb107c585017-kube-api-access-g5p47" (OuterVolumeSpecName: "kube-api-access-g5p47") pod "ed8f02d7-cd7c-4b04-9760-fb107c585017" (UID: "ed8f02d7-cd7c-4b04-9760-fb107c585017"). InnerVolumeSpecName "kube-api-access-g5p47". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.501655 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-scripts" (OuterVolumeSpecName: "scripts") pod "ed8f02d7-cd7c-4b04-9760-fb107c585017" (UID: "ed8f02d7-cd7c-4b04-9760-fb107c585017"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.522085 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ed8f02d7-cd7c-4b04-9760-fb107c585017" (UID: "ed8f02d7-cd7c-4b04-9760-fb107c585017"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.586301 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed8f02d7-cd7c-4b04-9760-fb107c585017-run-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.586331 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.586343 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.586352 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5p47\" (UniqueName: \"kubernetes.io/projected/ed8f02d7-cd7c-4b04-9760-fb107c585017-kube-api-access-g5p47\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.586360 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ed8f02d7-cd7c-4b04-9760-fb107c585017-log-httpd\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.598746 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ed8f02d7-cd7c-4b04-9760-fb107c585017" (UID: "ed8f02d7-cd7c-4b04-9760-fb107c585017"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.629533 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed8f02d7-cd7c-4b04-9760-fb107c585017" (UID: "ed8f02d7-cd7c-4b04-9760-fb107c585017"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.683389 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-config-data" (OuterVolumeSpecName: "config-data") pod "ed8f02d7-cd7c-4b04-9760-fb107c585017" (UID: "ed8f02d7-cd7c-4b04-9760-fb107c585017"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.689614 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.689648 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.689659 4872 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ed8f02d7-cd7c-4b04-9760-fb107c585017-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.708167 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.823400 4872 generic.go:334] "Generic (PLEG): container finished" podID="0855d9b8-26d6-4175-beed-a85fecaa3d29" containerID="ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6" exitCode=0 Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.823470 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" event={"ID":"0855d9b8-26d6-4175-beed-a85fecaa3d29","Type":"ContainerDied","Data":"ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6"} Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.823549 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" event={"ID":"0855d9b8-26d6-4175-beed-a85fecaa3d29","Type":"ContainerDied","Data":"837150f47c0d7e22121774811d91a1ddf01517ab70b9857384633428fc2efe57"} Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.823546 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-gthgc" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.823579 4872 scope.go:117] "RemoveContainer" containerID="ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.833120 4872 generic.go:334] "Generic (PLEG): container finished" podID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerID="ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b" exitCode=0 Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.834243 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.841140 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed8f02d7-cd7c-4b04-9760-fb107c585017","Type":"ContainerDied","Data":"ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b"} Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.841187 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ed8f02d7-cd7c-4b04-9760-fb107c585017","Type":"ContainerDied","Data":"8bd7e6fac1d13bf2be8686e2474a14ca44e69e8421d6d1046c07a198bcb0ffe6"} Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.852407 4872 scope.go:117] "RemoveContainer" containerID="8478f5e2ed375944270c60574080287bacc7b4055e3a7296d102e3d2d1571433" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.887597 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.892551 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q92zr\" (UniqueName: \"kubernetes.io/projected/0855d9b8-26d6-4175-beed-a85fecaa3d29-kube-api-access-q92zr\") pod \"0855d9b8-26d6-4175-beed-a85fecaa3d29\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.892835 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-dns-svc\") pod \"0855d9b8-26d6-4175-beed-a85fecaa3d29\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.892967 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-config\") pod \"0855d9b8-26d6-4175-beed-a85fecaa3d29\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.893061 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-ovsdbserver-nb\") pod \"0855d9b8-26d6-4175-beed-a85fecaa3d29\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.893136 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-ovsdbserver-sb\") pod \"0855d9b8-26d6-4175-beed-a85fecaa3d29\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.893277 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-dns-swift-storage-0\") pod \"0855d9b8-26d6-4175-beed-a85fecaa3d29\" (UID: \"0855d9b8-26d6-4175-beed-a85fecaa3d29\") " Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.894032 4872 scope.go:117] "RemoveContainer" containerID="ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6" Jan 26 09:26:44 crc kubenswrapper[4872]: E0126 09:26:44.895206 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6\": container with ID starting with ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6 not found: ID does not exist" containerID="ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.895347 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6"} err="failed to get container status \"ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6\": rpc error: code = NotFound desc = could not find container \"ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6\": container with ID starting with ee6825871cc55ce57a94040621b0e1681cc6eb2f4175501f1035d855df4d26f6 not found: ID does not exist" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.895389 4872 scope.go:117] "RemoveContainer" containerID="8478f5e2ed375944270c60574080287bacc7b4055e3a7296d102e3d2d1571433" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.907083 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0855d9b8-26d6-4175-beed-a85fecaa3d29-kube-api-access-q92zr" (OuterVolumeSpecName: "kube-api-access-q92zr") pod "0855d9b8-26d6-4175-beed-a85fecaa3d29" (UID: "0855d9b8-26d6-4175-beed-a85fecaa3d29"). InnerVolumeSpecName "kube-api-access-q92zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:44 crc kubenswrapper[4872]: E0126 09:26:44.907235 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8478f5e2ed375944270c60574080287bacc7b4055e3a7296d102e3d2d1571433\": container with ID starting with 8478f5e2ed375944270c60574080287bacc7b4055e3a7296d102e3d2d1571433 not found: ID does not exist" containerID="8478f5e2ed375944270c60574080287bacc7b4055e3a7296d102e3d2d1571433" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.907284 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8478f5e2ed375944270c60574080287bacc7b4055e3a7296d102e3d2d1571433"} err="failed to get container status \"8478f5e2ed375944270c60574080287bacc7b4055e3a7296d102e3d2d1571433\": rpc error: code = NotFound desc = could not find container \"8478f5e2ed375944270c60574080287bacc7b4055e3a7296d102e3d2d1571433\": container with ID starting with 8478f5e2ed375944270c60574080287bacc7b4055e3a7296d102e3d2d1571433 not found: ID does not exist" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.907315 4872 scope.go:117] "RemoveContainer" containerID="5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.907484 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.930644 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:26:44 crc kubenswrapper[4872]: E0126 09:26:44.931435 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0855d9b8-26d6-4175-beed-a85fecaa3d29" containerName="dnsmasq-dns" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.931533 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0855d9b8-26d6-4175-beed-a85fecaa3d29" containerName="dnsmasq-dns" Jan 26 09:26:44 crc kubenswrapper[4872]: E0126 09:26:44.931622 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="ceilometer-notification-agent" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.931678 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="ceilometer-notification-agent" Jan 26 09:26:44 crc kubenswrapper[4872]: E0126 09:26:44.931752 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="proxy-httpd" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.931819 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="proxy-httpd" Jan 26 09:26:44 crc kubenswrapper[4872]: E0126 09:26:44.931879 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="sg-core" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.931929 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="sg-core" Jan 26 09:26:44 crc kubenswrapper[4872]: E0126 09:26:44.931990 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0855d9b8-26d6-4175-beed-a85fecaa3d29" containerName="init" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.932054 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0855d9b8-26d6-4175-beed-a85fecaa3d29" containerName="init" Jan 26 09:26:44 crc kubenswrapper[4872]: E0126 09:26:44.932121 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="ceilometer-central-agent" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.932173 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="ceilometer-central-agent" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.932398 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="sg-core" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.932464 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="ceilometer-notification-agent" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.932534 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="0855d9b8-26d6-4175-beed-a85fecaa3d29" containerName="dnsmasq-dns" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.932592 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="ceilometer-central-agent" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.932649 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" containerName="proxy-httpd" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.935001 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.935059 4872 scope.go:117] "RemoveContainer" containerID="8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.945004 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.945409 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.945889 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.953572 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.972728 4872 scope.go:117] "RemoveContainer" containerID="65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.975238 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0855d9b8-26d6-4175-beed-a85fecaa3d29" (UID: "0855d9b8-26d6-4175-beed-a85fecaa3d29"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.981685 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-rmm5v"] Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.981928 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0855d9b8-26d6-4175-beed-a85fecaa3d29" (UID: "0855d9b8-26d6-4175-beed-a85fecaa3d29"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.989021 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-config" (OuterVolumeSpecName: "config") pod "0855d9b8-26d6-4175-beed-a85fecaa3d29" (UID: "0855d9b8-26d6-4175-beed-a85fecaa3d29"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:26:44 crc kubenswrapper[4872]: W0126 09:26:44.995421 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3eec5e98_7d8f_4ee7_9bb8_e6824f931b01.slice/crio-edfc90554d9c827052088642fb3de42c321778857402c964c84125c95ca29720 WatchSource:0}: Error finding container edfc90554d9c827052088642fb3de42c321778857402c964c84125c95ca29720: Status 404 returned error can't find the container with id edfc90554d9c827052088642fb3de42c321778857402c964c84125c95ca29720 Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.996332 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.996365 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q92zr\" (UniqueName: \"kubernetes.io/projected/0855d9b8-26d6-4175-beed-a85fecaa3d29-kube-api-access-q92zr\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.996378 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-dns-svc\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:44 crc kubenswrapper[4872]: I0126 09:26:44.996437 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-config\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.008075 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0855d9b8-26d6-4175-beed-a85fecaa3d29" (UID: "0855d9b8-26d6-4175-beed-a85fecaa3d29"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.009063 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0855d9b8-26d6-4175-beed-a85fecaa3d29" (UID: "0855d9b8-26d6-4175-beed-a85fecaa3d29"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.009101 4872 scope.go:117] "RemoveContainer" containerID="ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.039616 4872 scope.go:117] "RemoveContainer" containerID="5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4" Jan 26 09:26:45 crc kubenswrapper[4872]: E0126 09:26:45.040619 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4\": container with ID starting with 5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4 not found: ID does not exist" containerID="5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.040703 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4"} err="failed to get container status \"5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4\": rpc error: code = NotFound desc = could not find container \"5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4\": container with ID starting with 5f36bebcfcf2810f60f9ba035326ba7057b3b61ed7e7fe2f01982b1eddad54c4 not found: ID does not exist" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.040743 4872 scope.go:117] "RemoveContainer" containerID="8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9" Jan 26 09:26:45 crc kubenswrapper[4872]: E0126 09:26:45.041394 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9\": container with ID starting with 8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9 not found: ID does not exist" containerID="8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.041430 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9"} err="failed to get container status \"8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9\": rpc error: code = NotFound desc = could not find container \"8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9\": container with ID starting with 8d0500cced7cad621681a9653397a53a7066a91a3f896c8a5a71b93707c9c0e9 not found: ID does not exist" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.041450 4872 scope.go:117] "RemoveContainer" containerID="65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44" Jan 26 09:26:45 crc kubenswrapper[4872]: E0126 09:26:45.041722 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44\": container with ID starting with 65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44 not found: ID does not exist" containerID="65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.041743 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44"} err="failed to get container status \"65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44\": rpc error: code = NotFound desc = could not find container \"65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44\": container with ID starting with 65bf7dd54fe8353c4d544bc51e7d3a12dbb6fb9ed7ef71c3159c304919670b44 not found: ID does not exist" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.041760 4872 scope.go:117] "RemoveContainer" containerID="ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b" Jan 26 09:26:45 crc kubenswrapper[4872]: E0126 09:26:45.042022 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b\": container with ID starting with ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b not found: ID does not exist" containerID="ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.042052 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b"} err="failed to get container status \"ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b\": rpc error: code = NotFound desc = could not find container \"ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b\": container with ID starting with ba422534bf758b35998db77fbdda7bfb298dfbb34d96322556e7c69291bbde3b not found: ID does not exist" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.098531 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f535322c-098d-4f83-b91b-8cd72dd3339b-run-httpd\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.099064 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f535322c-098d-4f83-b91b-8cd72dd3339b-log-httpd\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.099114 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-config-data\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.099165 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.099311 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.099356 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9v6v\" (UniqueName: \"kubernetes.io/projected/f535322c-098d-4f83-b91b-8cd72dd3339b-kube-api-access-g9v6v\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.099443 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-scripts\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.099500 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.099859 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.099876 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0855d9b8-26d6-4175-beed-a85fecaa3d29-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.195370 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed8f02d7-cd7c-4b04-9760-fb107c585017" path="/var/lib/kubelet/pods/ed8f02d7-cd7c-4b04-9760-fb107c585017/volumes" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.202298 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f535322c-098d-4f83-b91b-8cd72dd3339b-log-httpd\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.202343 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f535322c-098d-4f83-b91b-8cd72dd3339b-run-httpd\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.202396 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-config-data\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.202464 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.202960 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f535322c-098d-4f83-b91b-8cd72dd3339b-log-httpd\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.203011 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f535322c-098d-4f83-b91b-8cd72dd3339b-run-httpd\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.203489 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.203606 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9v6v\" (UniqueName: \"kubernetes.io/projected/f535322c-098d-4f83-b91b-8cd72dd3339b-kube-api-access-g9v6v\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.203679 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-scripts\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.203728 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.207334 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.209003 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-config-data\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.211012 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-scripts\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.210349 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.214702 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f535322c-098d-4f83-b91b-8cd72dd3339b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.229390 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9v6v\" (UniqueName: \"kubernetes.io/projected/f535322c-098d-4f83-b91b-8cd72dd3339b-kube-api-access-g9v6v\") pod \"ceilometer-0\" (UID: \"f535322c-098d-4f83-b91b-8cd72dd3339b\") " pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.276297 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.400061 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-gthgc"] Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.407868 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-gthgc"] Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.798764 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.847177 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rmm5v" event={"ID":"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01","Type":"ContainerStarted","Data":"0ccd9721504662de5f52f20d2d968d24d64df604f11270c047c06af15d559bc8"} Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.847245 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rmm5v" event={"ID":"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01","Type":"ContainerStarted","Data":"edfc90554d9c827052088642fb3de42c321778857402c964c84125c95ca29720"} Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.849662 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f535322c-098d-4f83-b91b-8cd72dd3339b","Type":"ContainerStarted","Data":"bd9512114983c4a9d437ba010e4ced159a816f659695f79befb28e71fe27d102"} Jan 26 09:26:45 crc kubenswrapper[4872]: I0126 09:26:45.878782 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-rmm5v" podStartSLOduration=1.878757597 podStartE2EDuration="1.878757597s" podCreationTimestamp="2026-01-26 09:26:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:26:45.863924405 +0000 UTC m=+1139.172764206" watchObservedRunningTime="2026-01-26 09:26:45.878757597 +0000 UTC m=+1139.187597398" Jan 26 09:26:46 crc kubenswrapper[4872]: I0126 09:26:46.866755 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f535322c-098d-4f83-b91b-8cd72dd3339b","Type":"ContainerStarted","Data":"fa26b4d7b0dda215bb00274bdb243a51c90d79896acb6eb27c2447e700cefea4"} Jan 26 09:26:47 crc kubenswrapper[4872]: I0126 09:26:47.198370 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0855d9b8-26d6-4175-beed-a85fecaa3d29" path="/var/lib/kubelet/pods/0855d9b8-26d6-4175-beed-a85fecaa3d29/volumes" Jan 26 09:26:47 crc kubenswrapper[4872]: I0126 09:26:47.880553 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f535322c-098d-4f83-b91b-8cd72dd3339b","Type":"ContainerStarted","Data":"d90fc8ede34f7bf6948e4709fe4dd849d59a2d4f654ce4578a25b59901af8cea"} Jan 26 09:26:48 crc kubenswrapper[4872]: I0126 09:26:48.907853 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f535322c-098d-4f83-b91b-8cd72dd3339b","Type":"ContainerStarted","Data":"518af3337ea49fc340eb26d225dfd8ae4b895c79776305b8f41b5190a52a8856"} Jan 26 09:26:50 crc kubenswrapper[4872]: I0126 09:26:50.931634 4872 generic.go:334] "Generic (PLEG): container finished" podID="3eec5e98-7d8f-4ee7-9bb8-e6824f931b01" containerID="0ccd9721504662de5f52f20d2d968d24d64df604f11270c047c06af15d559bc8" exitCode=0 Jan 26 09:26:50 crc kubenswrapper[4872]: I0126 09:26:50.932427 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rmm5v" event={"ID":"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01","Type":"ContainerDied","Data":"0ccd9721504662de5f52f20d2d968d24d64df604f11270c047c06af15d559bc8"} Jan 26 09:26:50 crc kubenswrapper[4872]: I0126 09:26:50.936864 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f535322c-098d-4f83-b91b-8cd72dd3339b","Type":"ContainerStarted","Data":"47b92ef85fff4338d8becff13a509d0eba113f650bb4900c4fc8f658fb94fd95"} Jan 26 09:26:50 crc kubenswrapper[4872]: I0126 09:26:50.937106 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Jan 26 09:26:51 crc kubenswrapper[4872]: I0126 09:26:51.004315 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.174277949 podStartE2EDuration="7.004283955s" podCreationTimestamp="2026-01-26 09:26:44 +0000 UTC" firstStartedPulling="2026-01-26 09:26:45.797995413 +0000 UTC m=+1139.106835214" lastFinishedPulling="2026-01-26 09:26:49.628001409 +0000 UTC m=+1142.936841220" observedRunningTime="2026-01-26 09:26:50.991570199 +0000 UTC m=+1144.300410050" watchObservedRunningTime="2026-01-26 09:26:51.004283955 +0000 UTC m=+1144.313123776" Jan 26 09:26:51 crc kubenswrapper[4872]: I0126 09:26:51.194915 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 09:26:51 crc kubenswrapper[4872]: I0126 09:26:51.194966 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.213794 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.212291 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.438326 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.520949 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-combined-ca-bundle\") pod \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.521095 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hb5hw\" (UniqueName: \"kubernetes.io/projected/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-kube-api-access-hb5hw\") pod \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.521131 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-config-data\") pod \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.521297 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-scripts\") pod \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\" (UID: \"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01\") " Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.534027 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-kube-api-access-hb5hw" (OuterVolumeSpecName: "kube-api-access-hb5hw") pod "3eec5e98-7d8f-4ee7-9bb8-e6824f931b01" (UID: "3eec5e98-7d8f-4ee7-9bb8-e6824f931b01"). InnerVolumeSpecName "kube-api-access-hb5hw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.536013 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-scripts" (OuterVolumeSpecName: "scripts") pod "3eec5e98-7d8f-4ee7-9bb8-e6824f931b01" (UID: "3eec5e98-7d8f-4ee7-9bb8-e6824f931b01"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.557000 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3eec5e98-7d8f-4ee7-9bb8-e6824f931b01" (UID: "3eec5e98-7d8f-4ee7-9bb8-e6824f931b01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.558449 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-config-data" (OuterVolumeSpecName: "config-data") pod "3eec5e98-7d8f-4ee7-9bb8-e6824f931b01" (UID: "3eec5e98-7d8f-4ee7-9bb8-e6824f931b01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.624136 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.624205 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hb5hw\" (UniqueName: \"kubernetes.io/projected/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-kube-api-access-hb5hw\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.624240 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.624263 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01-scripts\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.968471 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-rmm5v" event={"ID":"3eec5e98-7d8f-4ee7-9bb8-e6824f931b01","Type":"ContainerDied","Data":"edfc90554d9c827052088642fb3de42c321778857402c964c84125c95ca29720"} Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.968530 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edfc90554d9c827052088642fb3de42c321778857402c964c84125c95ca29720" Jan 26 09:26:52 crc kubenswrapper[4872]: I0126 09:26:52.969082 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-rmm5v" Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.249471 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.249528 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.249854 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" containerName="nova-api-log" containerID="cri-o://52eb3db4a582c6797b9128c5130fd02f64f645682d724efd67e13f4932758626" gracePeriod=30 Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.250136 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" containerName="nova-api-api" containerID="cri-o://44b717b6cbdce542f10ea68ad34a469f7fea7c2ac580e2d6797e0d34a3464b6a" gracePeriod=30 Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.250521 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="19b7b80d-07f5-472b-b3df-dce1de4223a6" containerName="nova-scheduler-scheduler" containerID="cri-o://44e0355a8eff3442560d63c49f4c428c6bef0da33f9cfdb0fb7c6e586980d39c" gracePeriod=30 Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.280021 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.280574 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerName="nova-metadata-log" containerID="cri-o://569eedd9c241b70919ce97797c49d587e23e0812b018b6bb6b13d967dbfd30e3" gracePeriod=30 Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.281112 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerName="nova-metadata-metadata" containerID="cri-o://74d11d13b2a5c9ccf88d535de057db9ed169578d1016a5efdf2b773f425a746b" gracePeriod=30 Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.340003 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-conductor-0" Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.984381 4872 generic.go:334] "Generic (PLEG): container finished" podID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" containerID="52eb3db4a582c6797b9128c5130fd02f64f645682d724efd67e13f4932758626" exitCode=143 Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.984536 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55","Type":"ContainerDied","Data":"52eb3db4a582c6797b9128c5130fd02f64f645682d724efd67e13f4932758626"} Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.986631 4872 generic.go:334] "Generic (PLEG): container finished" podID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerID="569eedd9c241b70919ce97797c49d587e23e0812b018b6bb6b13d967dbfd30e3" exitCode=143 Jan 26 09:26:53 crc kubenswrapper[4872]: I0126 09:26:53.986670 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a4c81032-5e90-4b7d-a27d-a46b7e052269","Type":"ContainerDied","Data":"569eedd9c241b70919ce97797c49d587e23e0812b018b6bb6b13d967dbfd30e3"} Jan 26 09:26:55 crc kubenswrapper[4872]: I0126 09:26:55.006373 4872 generic.go:334] "Generic (PLEG): container finished" podID="19b7b80d-07f5-472b-b3df-dce1de4223a6" containerID="44e0355a8eff3442560d63c49f4c428c6bef0da33f9cfdb0fb7c6e586980d39c" exitCode=0 Jan 26 09:26:55 crc kubenswrapper[4872]: I0126 09:26:55.006532 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19b7b80d-07f5-472b-b3df-dce1de4223a6","Type":"ContainerDied","Data":"44e0355a8eff3442560d63c49f4c428c6bef0da33f9cfdb0fb7c6e586980d39c"} Jan 26 09:26:55 crc kubenswrapper[4872]: I0126 09:26:55.178375 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 09:26:55 crc kubenswrapper[4872]: I0126 09:26:55.317767 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19b7b80d-07f5-472b-b3df-dce1de4223a6-config-data\") pod \"19b7b80d-07f5-472b-b3df-dce1de4223a6\" (UID: \"19b7b80d-07f5-472b-b3df-dce1de4223a6\") " Jan 26 09:26:55 crc kubenswrapper[4872]: I0126 09:26:55.317873 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnc69\" (UniqueName: \"kubernetes.io/projected/19b7b80d-07f5-472b-b3df-dce1de4223a6-kube-api-access-hnc69\") pod \"19b7b80d-07f5-472b-b3df-dce1de4223a6\" (UID: \"19b7b80d-07f5-472b-b3df-dce1de4223a6\") " Jan 26 09:26:55 crc kubenswrapper[4872]: I0126 09:26:55.318054 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b7b80d-07f5-472b-b3df-dce1de4223a6-combined-ca-bundle\") pod \"19b7b80d-07f5-472b-b3df-dce1de4223a6\" (UID: \"19b7b80d-07f5-472b-b3df-dce1de4223a6\") " Jan 26 09:26:55 crc kubenswrapper[4872]: I0126 09:26:55.329421 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19b7b80d-07f5-472b-b3df-dce1de4223a6-kube-api-access-hnc69" (OuterVolumeSpecName: "kube-api-access-hnc69") pod "19b7b80d-07f5-472b-b3df-dce1de4223a6" (UID: "19b7b80d-07f5-472b-b3df-dce1de4223a6"). InnerVolumeSpecName "kube-api-access-hnc69". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:55 crc kubenswrapper[4872]: I0126 09:26:55.356292 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b7b80d-07f5-472b-b3df-dce1de4223a6-config-data" (OuterVolumeSpecName: "config-data") pod "19b7b80d-07f5-472b-b3df-dce1de4223a6" (UID: "19b7b80d-07f5-472b-b3df-dce1de4223a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:55 crc kubenswrapper[4872]: I0126 09:26:55.365174 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19b7b80d-07f5-472b-b3df-dce1de4223a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "19b7b80d-07f5-472b-b3df-dce1de4223a6" (UID: "19b7b80d-07f5-472b-b3df-dce1de4223a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:55 crc kubenswrapper[4872]: I0126 09:26:55.422590 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/19b7b80d-07f5-472b-b3df-dce1de4223a6-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:55 crc kubenswrapper[4872]: I0126 09:26:55.422644 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnc69\" (UniqueName: \"kubernetes.io/projected/19b7b80d-07f5-472b-b3df-dce1de4223a6-kube-api-access-hnc69\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:55 crc kubenswrapper[4872]: I0126 09:26:55.422668 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19b7b80d-07f5-472b-b3df-dce1de4223a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.024751 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"19b7b80d-07f5-472b-b3df-dce1de4223a6","Type":"ContainerDied","Data":"97e4826b92985e91c7d106f133a209b5d2aea078234c3f8911a9f1d281ff7093"} Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.024893 4872 scope.go:117] "RemoveContainer" containerID="44e0355a8eff3442560d63c49f4c428c6bef0da33f9cfdb0fb7c6e586980d39c" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.024895 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.071424 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.087994 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.113719 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:26:56 crc kubenswrapper[4872]: E0126 09:26:56.116386 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eec5e98-7d8f-4ee7-9bb8-e6824f931b01" containerName="nova-manage" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.116438 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eec5e98-7d8f-4ee7-9bb8-e6824f931b01" containerName="nova-manage" Jan 26 09:26:56 crc kubenswrapper[4872]: E0126 09:26:56.116454 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19b7b80d-07f5-472b-b3df-dce1de4223a6" containerName="nova-scheduler-scheduler" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.116463 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="19b7b80d-07f5-472b-b3df-dce1de4223a6" containerName="nova-scheduler-scheduler" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.116697 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="19b7b80d-07f5-472b-b3df-dce1de4223a6" containerName="nova-scheduler-scheduler" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.116715 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3eec5e98-7d8f-4ee7-9bb8-e6824f931b01" containerName="nova-manage" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.117631 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.122028 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.127003 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.240448 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr7xs\" (UniqueName: \"kubernetes.io/projected/c97bb470-8700-44dc-99a4-4e39b14aa332-kube-api-access-rr7xs\") pod \"nova-scheduler-0\" (UID: \"c97bb470-8700-44dc-99a4-4e39b14aa332\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.241011 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c97bb470-8700-44dc-99a4-4e39b14aa332-config-data\") pod \"nova-scheduler-0\" (UID: \"c97bb470-8700-44dc-99a4-4e39b14aa332\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.241075 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97bb470-8700-44dc-99a4-4e39b14aa332-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c97bb470-8700-44dc-99a4-4e39b14aa332\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.343418 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr7xs\" (UniqueName: \"kubernetes.io/projected/c97bb470-8700-44dc-99a4-4e39b14aa332-kube-api-access-rr7xs\") pod \"nova-scheduler-0\" (UID: \"c97bb470-8700-44dc-99a4-4e39b14aa332\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.343546 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c97bb470-8700-44dc-99a4-4e39b14aa332-config-data\") pod \"nova-scheduler-0\" (UID: \"c97bb470-8700-44dc-99a4-4e39b14aa332\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.343590 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97bb470-8700-44dc-99a4-4e39b14aa332-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c97bb470-8700-44dc-99a4-4e39b14aa332\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.354426 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c97bb470-8700-44dc-99a4-4e39b14aa332-config-data\") pod \"nova-scheduler-0\" (UID: \"c97bb470-8700-44dc-99a4-4e39b14aa332\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.357390 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c97bb470-8700-44dc-99a4-4e39b14aa332-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c97bb470-8700-44dc-99a4-4e39b14aa332\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.364118 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr7xs\" (UniqueName: \"kubernetes.io/projected/c97bb470-8700-44dc-99a4-4e39b14aa332-kube-api-access-rr7xs\") pod \"nova-scheduler-0\" (UID: \"c97bb470-8700-44dc-99a4-4e39b14aa332\") " pod="openstack/nova-scheduler-0" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.454792 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.501692 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:37848->10.217.0.202:8775: read: connection reset by peer" Jan 26 09:26:56 crc kubenswrapper[4872]: I0126 09:26:56.501719 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:37850->10.217.0.202:8775: read: connection reset by peer" Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.020744 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.042939 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c97bb470-8700-44dc-99a4-4e39b14aa332","Type":"ContainerStarted","Data":"c368b7047c4c9bcbe85affc708ee0a537c8354e998feb4d5ede083c6284e9e33"} Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.048239 4872 generic.go:334] "Generic (PLEG): container finished" podID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerID="74d11d13b2a5c9ccf88d535de057db9ed169578d1016a5efdf2b773f425a746b" exitCode=0 Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.048302 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a4c81032-5e90-4b7d-a27d-a46b7e052269","Type":"ContainerDied","Data":"74d11d13b2a5c9ccf88d535de057db9ed169578d1016a5efdf2b773f425a746b"} Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.051275 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.061848 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-combined-ca-bundle\") pod \"a4c81032-5e90-4b7d-a27d-a46b7e052269\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.061978 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-nova-metadata-tls-certs\") pod \"a4c81032-5e90-4b7d-a27d-a46b7e052269\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.062085 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4vr6\" (UniqueName: \"kubernetes.io/projected/a4c81032-5e90-4b7d-a27d-a46b7e052269-kube-api-access-t4vr6\") pod \"a4c81032-5e90-4b7d-a27d-a46b7e052269\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.062202 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-config-data\") pod \"a4c81032-5e90-4b7d-a27d-a46b7e052269\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.062261 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4c81032-5e90-4b7d-a27d-a46b7e052269-logs\") pod \"a4c81032-5e90-4b7d-a27d-a46b7e052269\" (UID: \"a4c81032-5e90-4b7d-a27d-a46b7e052269\") " Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.063343 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4c81032-5e90-4b7d-a27d-a46b7e052269-logs" (OuterVolumeSpecName: "logs") pod "a4c81032-5e90-4b7d-a27d-a46b7e052269" (UID: "a4c81032-5e90-4b7d-a27d-a46b7e052269"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.068741 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4c81032-5e90-4b7d-a27d-a46b7e052269-kube-api-access-t4vr6" (OuterVolumeSpecName: "kube-api-access-t4vr6") pod "a4c81032-5e90-4b7d-a27d-a46b7e052269" (UID: "a4c81032-5e90-4b7d-a27d-a46b7e052269"). InnerVolumeSpecName "kube-api-access-t4vr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.108993 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-config-data" (OuterVolumeSpecName: "config-data") pod "a4c81032-5e90-4b7d-a27d-a46b7e052269" (UID: "a4c81032-5e90-4b7d-a27d-a46b7e052269"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.135550 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4c81032-5e90-4b7d-a27d-a46b7e052269" (UID: "a4c81032-5e90-4b7d-a27d-a46b7e052269"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.142023 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "a4c81032-5e90-4b7d-a27d-a46b7e052269" (UID: "a4c81032-5e90-4b7d-a27d-a46b7e052269"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.164835 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.164935 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4c81032-5e90-4b7d-a27d-a46b7e052269-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.164946 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.164962 4872 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4c81032-5e90-4b7d-a27d-a46b7e052269-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.164976 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4vr6\" (UniqueName: \"kubernetes.io/projected/a4c81032-5e90-4b7d-a27d-a46b7e052269-kube-api-access-t4vr6\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:57 crc kubenswrapper[4872]: I0126 09:26:57.201933 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19b7b80d-07f5-472b-b3df-dce1de4223a6" path="/var/lib/kubelet/pods/19b7b80d-07f5-472b-b3df-dce1de4223a6/volumes" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.062440 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a4c81032-5e90-4b7d-a27d-a46b7e052269","Type":"ContainerDied","Data":"89c409f1221aa4dbfd7a0a9a598a09f35e2aa0e0baf96cb357e2210e17b3b558"} Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.062968 4872 scope.go:117] "RemoveContainer" containerID="74d11d13b2a5c9ccf88d535de057db9ed169578d1016a5efdf2b773f425a746b" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.062504 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.067523 4872 generic.go:334] "Generic (PLEG): container finished" podID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" containerID="44b717b6cbdce542f10ea68ad34a469f7fea7c2ac580e2d6797e0d34a3464b6a" exitCode=0 Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.067596 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55","Type":"ContainerDied","Data":"44b717b6cbdce542f10ea68ad34a469f7fea7c2ac580e2d6797e0d34a3464b6a"} Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.069788 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c97bb470-8700-44dc-99a4-4e39b14aa332","Type":"ContainerStarted","Data":"02c29a4f6b9f8e4f0dc54d56da2e1ba6e35669bb32696a6506904747720d47a5"} Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.095762 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.109973 4872 scope.go:117] "RemoveContainer" containerID="569eedd9c241b70919ce97797c49d587e23e0812b018b6bb6b13d967dbfd30e3" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.109970 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.113400 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.113362011 podStartE2EDuration="2.113362011s" podCreationTimestamp="2026-01-26 09:26:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:26:58.102196155 +0000 UTC m=+1151.411036016" watchObservedRunningTime="2026-01-26 09:26:58.113362011 +0000 UTC m=+1151.422201812" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.142911 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:58 crc kubenswrapper[4872]: E0126 09:26:58.143485 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerName="nova-metadata-metadata" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.143514 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerName="nova-metadata-metadata" Jan 26 09:26:58 crc kubenswrapper[4872]: E0126 09:26:58.143545 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerName="nova-metadata-log" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.143554 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerName="nova-metadata-log" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.143757 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerName="nova-metadata-log" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.143791 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" containerName="nova-metadata-metadata" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.144903 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.150477 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.150722 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.153580 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.217116 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.289763 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/735c976b-3cf3-4b78-b5fc-4d92d09f410c-config-data\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.289925 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/735c976b-3cf3-4b78-b5fc-4d92d09f410c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.289969 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/735c976b-3cf3-4b78-b5fc-4d92d09f410c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.290041 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/735c976b-3cf3-4b78-b5fc-4d92d09f410c-logs\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.290078 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rzf7\" (UniqueName: \"kubernetes.io/projected/735c976b-3cf3-4b78-b5fc-4d92d09f410c-kube-api-access-2rzf7\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.390828 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-combined-ca-bundle\") pod \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.391226 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-internal-tls-certs\") pod \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.391258 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-public-tls-certs\") pod \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.391318 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6s762\" (UniqueName: \"kubernetes.io/projected/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-kube-api-access-6s762\") pod \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.391392 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-logs\") pod \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.391442 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-config-data\") pod \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\" (UID: \"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55\") " Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.391763 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/735c976b-3cf3-4b78-b5fc-4d92d09f410c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.391827 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/735c976b-3cf3-4b78-b5fc-4d92d09f410c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.391890 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/735c976b-3cf3-4b78-b5fc-4d92d09f410c-logs\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.391934 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rzf7\" (UniqueName: \"kubernetes.io/projected/735c976b-3cf3-4b78-b5fc-4d92d09f410c-kube-api-access-2rzf7\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.392001 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/735c976b-3cf3-4b78-b5fc-4d92d09f410c-config-data\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.392404 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-logs" (OuterVolumeSpecName: "logs") pod "f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" (UID: "f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.392592 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/735c976b-3cf3-4b78-b5fc-4d92d09f410c-logs\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.398301 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/735c976b-3cf3-4b78-b5fc-4d92d09f410c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.401038 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-kube-api-access-6s762" (OuterVolumeSpecName: "kube-api-access-6s762") pod "f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" (UID: "f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55"). InnerVolumeSpecName "kube-api-access-6s762". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.405473 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/735c976b-3cf3-4b78-b5fc-4d92d09f410c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.405787 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/735c976b-3cf3-4b78-b5fc-4d92d09f410c-config-data\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.408530 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rzf7\" (UniqueName: \"kubernetes.io/projected/735c976b-3cf3-4b78-b5fc-4d92d09f410c-kube-api-access-2rzf7\") pod \"nova-metadata-0\" (UID: \"735c976b-3cf3-4b78-b5fc-4d92d09f410c\") " pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.422771 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" (UID: "f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.433445 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-config-data" (OuterVolumeSpecName: "config-data") pod "f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" (UID: "f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.446518 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" (UID: "f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.448894 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" (UID: "f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.494677 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.494718 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.494735 4872 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.494745 4872 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-public-tls-certs\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.494756 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6s762\" (UniqueName: \"kubernetes.io/projected/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-kube-api-access-6s762\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.494766 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55-logs\") on node \"crc\" DevicePath \"\"" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.530769 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Jan 26 09:26:58 crc kubenswrapper[4872]: I0126 09:26:58.860388 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.091233 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"735c976b-3cf3-4b78-b5fc-4d92d09f410c","Type":"ContainerStarted","Data":"7c061ecc8c2bc81e705212a889171a02042521d19758eb71c1e4169ab65177af"} Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.091288 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"735c976b-3cf3-4b78-b5fc-4d92d09f410c","Type":"ContainerStarted","Data":"a74045cdbbcd730fa3ee4d318acfac699e4cc1900b6bd0a6118e7529894c3385"} Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.097838 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.097930 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55","Type":"ContainerDied","Data":"e9062b2d32ee86f79034f94f8bbfb74c07bc83b645c6b099a962ffa6b0ce5787"} Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.098018 4872 scope.go:117] "RemoveContainer" containerID="44b717b6cbdce542f10ea68ad34a469f7fea7c2ac580e2d6797e0d34a3464b6a" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.125322 4872 scope.go:117] "RemoveContainer" containerID="52eb3db4a582c6797b9128c5130fd02f64f645682d724efd67e13f4932758626" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.140860 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.149702 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.170449 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:59 crc kubenswrapper[4872]: E0126 09:26:59.170966 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" containerName="nova-api-api" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.170989 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" containerName="nova-api-api" Jan 26 09:26:59 crc kubenswrapper[4872]: E0126 09:26:59.171010 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" containerName="nova-api-log" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.171020 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" containerName="nova-api-log" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.171206 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" containerName="nova-api-api" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.171229 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" containerName="nova-api-log" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.172261 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.174413 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.174466 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.181149 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.200224 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4c81032-5e90-4b7d-a27d-a46b7e052269" path="/var/lib/kubelet/pods/a4c81032-5e90-4b7d-a27d-a46b7e052269/volumes" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.201071 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55" path="/var/lib/kubelet/pods/f728fc3d-3a48-4c25-b2ac-6d1d6ef0aa55/volumes" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.222483 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.320976 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.321086 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hz42\" (UniqueName: \"kubernetes.io/projected/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-kube-api-access-7hz42\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.321155 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-public-tls-certs\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.321183 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-config-data\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.321421 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.321671 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-logs\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.424358 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-public-tls-certs\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.424428 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-config-data\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.424458 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.424518 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-logs\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.424599 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.424634 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hz42\" (UniqueName: \"kubernetes.io/projected/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-kube-api-access-7hz42\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.425947 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-logs\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.431296 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.433115 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-public-tls-certs\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.434028 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.438543 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-config-data\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.445362 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hz42\" (UniqueName: \"kubernetes.io/projected/12b5f34c-5cc6-448f-ac2f-420df97eaaa3-kube-api-access-7hz42\") pod \"nova-api-0\" (UID: \"12b5f34c-5cc6-448f-ac2f-420df97eaaa3\") " pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.488944 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Jan 26 09:26:59 crc kubenswrapper[4872]: I0126 09:26:59.984570 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Jan 26 09:27:00 crc kubenswrapper[4872]: I0126 09:27:00.127775 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"735c976b-3cf3-4b78-b5fc-4d92d09f410c","Type":"ContainerStarted","Data":"d4794413b0e55088b949d18f00339ab925a68e2bd567964c6d99ae65d8107fe6"} Jan 26 09:27:00 crc kubenswrapper[4872]: I0126 09:27:00.131195 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12b5f34c-5cc6-448f-ac2f-420df97eaaa3","Type":"ContainerStarted","Data":"7aad4df062279eddc4d7eb7da03d691f32d588573faff0133ebdc6fa944b6d9e"} Jan 26 09:27:00 crc kubenswrapper[4872]: I0126 09:27:00.165468 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.165443649 podStartE2EDuration="2.165443649s" podCreationTimestamp="2026-01-26 09:26:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:27:00.154736344 +0000 UTC m=+1153.463576145" watchObservedRunningTime="2026-01-26 09:27:00.165443649 +0000 UTC m=+1153.474283450" Jan 26 09:27:01 crc kubenswrapper[4872]: I0126 09:27:01.150406 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12b5f34c-5cc6-448f-ac2f-420df97eaaa3","Type":"ContainerStarted","Data":"54601fd054a904f9683eb59af3c34e47bc8960b38e22d5bf8baddd8ad6a7a938"} Jan 26 09:27:01 crc kubenswrapper[4872]: I0126 09:27:01.151003 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12b5f34c-5cc6-448f-ac2f-420df97eaaa3","Type":"ContainerStarted","Data":"cc9545fc2ff8a66ebfb7fbc2c50c38f59b484d4f151126fe1b9f055537758e8b"} Jan 26 09:27:01 crc kubenswrapper[4872]: I0126 09:27:01.182996 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.182942273 podStartE2EDuration="2.182942273s" podCreationTimestamp="2026-01-26 09:26:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:27:01.172433244 +0000 UTC m=+1154.481273045" watchObservedRunningTime="2026-01-26 09:27:01.182942273 +0000 UTC m=+1154.491782074" Jan 26 09:27:01 crc kubenswrapper[4872]: I0126 09:27:01.455320 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Jan 26 09:27:03 crc kubenswrapper[4872]: I0126 09:27:03.531470 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 09:27:03 crc kubenswrapper[4872]: I0126 09:27:03.531787 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Jan 26 09:27:06 crc kubenswrapper[4872]: I0126 09:27:06.455377 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Jan 26 09:27:06 crc kubenswrapper[4872]: I0126 09:27:06.492849 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Jan 26 09:27:07 crc kubenswrapper[4872]: I0126 09:27:07.293865 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Jan 26 09:27:08 crc kubenswrapper[4872]: I0126 09:27:08.534416 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 26 09:27:08 crc kubenswrapper[4872]: I0126 09:27:08.534790 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Jan 26 09:27:09 crc kubenswrapper[4872]: I0126 09:27:09.489891 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 09:27:09 crc kubenswrapper[4872]: I0126 09:27:09.490323 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Jan 26 09:27:09 crc kubenswrapper[4872]: I0126 09:27:09.543097 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="735c976b-3cf3-4b78-b5fc-4d92d09f410c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 09:27:09 crc kubenswrapper[4872]: I0126 09:27:09.543161 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="735c976b-3cf3-4b78-b5fc-4d92d09f410c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 09:27:10 crc kubenswrapper[4872]: I0126 09:27:10.227856 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:27:10 crc kubenswrapper[4872]: I0126 09:27:10.227966 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:27:10 crc kubenswrapper[4872]: I0126 09:27:10.509016 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="12b5f34c-5cc6-448f-ac2f-420df97eaaa3" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Jan 26 09:27:10 crc kubenswrapper[4872]: I0126 09:27:10.509051 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="12b5f34c-5cc6-448f-ac2f-420df97eaaa3" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 09:27:15 crc kubenswrapper[4872]: I0126 09:27:15.288026 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Jan 26 09:27:18 crc kubenswrapper[4872]: I0126 09:27:18.538518 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 26 09:27:18 crc kubenswrapper[4872]: I0126 09:27:18.545156 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Jan 26 09:27:18 crc kubenswrapper[4872]: I0126 09:27:18.545731 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 26 09:27:19 crc kubenswrapper[4872]: I0126 09:27:19.427377 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Jan 26 09:27:19 crc kubenswrapper[4872]: I0126 09:27:19.498953 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 26 09:27:19 crc kubenswrapper[4872]: I0126 09:27:19.499477 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 26 09:27:19 crc kubenswrapper[4872]: I0126 09:27:19.500841 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Jan 26 09:27:19 crc kubenswrapper[4872]: I0126 09:27:19.507198 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 26 09:27:20 crc kubenswrapper[4872]: I0126 09:27:20.426723 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Jan 26 09:27:20 crc kubenswrapper[4872]: I0126 09:27:20.438035 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Jan 26 09:27:28 crc kubenswrapper[4872]: I0126 09:27:28.831727 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 09:27:29 crc kubenswrapper[4872]: I0126 09:27:29.916137 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 09:27:33 crc kubenswrapper[4872]: I0126 09:27:33.719438 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="82c8133d-d299-4c99-b8a0-96099e834e8c" containerName="rabbitmq" containerID="cri-o://6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a" gracePeriod=604796 Jan 26 09:27:34 crc kubenswrapper[4872]: I0126 09:27:34.275727 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="82c8133d-d299-4c99-b8a0-96099e834e8c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.96:5671: connect: connection refused" Jan 26 09:27:34 crc kubenswrapper[4872]: I0126 09:27:34.562945 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="0792ad43-38af-4c94-bd3b-fc7d3a7af5da" containerName="rabbitmq" containerID="cri-o://2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5" gracePeriod=604796 Jan 26 09:27:34 crc kubenswrapper[4872]: I0126 09:27:34.613432 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="0792ad43-38af-4c94-bd3b-fc7d3a7af5da" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.227662 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.228331 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.228383 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.229022 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"806f909f0f07687dd6ea8e62ed9234f07922743fa23a5604fb5f62e36f63f8b0"} pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.229083 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" containerID="cri-o://806f909f0f07687dd6ea8e62ed9234f07922743fa23a5604fb5f62e36f63f8b0" gracePeriod=600 Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.463543 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.516359 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-confd\") pod \"82c8133d-d299-4c99-b8a0-96099e834e8c\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.516458 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-plugins\") pod \"82c8133d-d299-4c99-b8a0-96099e834e8c\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.516501 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-plugins-conf\") pod \"82c8133d-d299-4c99-b8a0-96099e834e8c\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.516606 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82c8133d-d299-4c99-b8a0-96099e834e8c-erlang-cookie-secret\") pod \"82c8133d-d299-4c99-b8a0-96099e834e8c\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.516721 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-erlang-cookie\") pod \"82c8133d-d299-4c99-b8a0-96099e834e8c\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.516765 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-tls\") pod \"82c8133d-d299-4c99-b8a0-96099e834e8c\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.516823 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82c8133d-d299-4c99-b8a0-96099e834e8c-pod-info\") pod \"82c8133d-d299-4c99-b8a0-96099e834e8c\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.516895 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"82c8133d-d299-4c99-b8a0-96099e834e8c\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.516994 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-server-conf\") pod \"82c8133d-d299-4c99-b8a0-96099e834e8c\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.517064 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw2k5\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-kube-api-access-xw2k5\") pod \"82c8133d-d299-4c99-b8a0-96099e834e8c\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.517110 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-config-data\") pod \"82c8133d-d299-4c99-b8a0-96099e834e8c\" (UID: \"82c8133d-d299-4c99-b8a0-96099e834e8c\") " Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.517766 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "82c8133d-d299-4c99-b8a0-96099e834e8c" (UID: "82c8133d-d299-4c99-b8a0-96099e834e8c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.518602 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "82c8133d-d299-4c99-b8a0-96099e834e8c" (UID: "82c8133d-d299-4c99-b8a0-96099e834e8c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.520235 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "82c8133d-d299-4c99-b8a0-96099e834e8c" (UID: "82c8133d-d299-4c99-b8a0-96099e834e8c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.526047 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "82c8133d-d299-4c99-b8a0-96099e834e8c" (UID: "82c8133d-d299-4c99-b8a0-96099e834e8c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.530991 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82c8133d-d299-4c99-b8a0-96099e834e8c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "82c8133d-d299-4c99-b8a0-96099e834e8c" (UID: "82c8133d-d299-4c99-b8a0-96099e834e8c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.531123 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/82c8133d-d299-4c99-b8a0-96099e834e8c-pod-info" (OuterVolumeSpecName: "pod-info") pod "82c8133d-d299-4c99-b8a0-96099e834e8c" (UID: "82c8133d-d299-4c99-b8a0-96099e834e8c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.531211 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-kube-api-access-xw2k5" (OuterVolumeSpecName: "kube-api-access-xw2k5") pod "82c8133d-d299-4c99-b8a0-96099e834e8c" (UID: "82c8133d-d299-4c99-b8a0-96099e834e8c"). InnerVolumeSpecName "kube-api-access-xw2k5". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.546759 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "82c8133d-d299-4c99-b8a0-96099e834e8c" (UID: "82c8133d-d299-4c99-b8a0-96099e834e8c"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.606398 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-config-data" (OuterVolumeSpecName: "config-data") pod "82c8133d-d299-4c99-b8a0-96099e834e8c" (UID: "82c8133d-d299-4c99-b8a0-96099e834e8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.619485 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.619519 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.619545 4872 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.619554 4872 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82c8133d-d299-4c99-b8a0-96099e834e8c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.619565 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.619573 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.619581 4872 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82c8133d-d299-4c99-b8a0-96099e834e8c-pod-info\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.619619 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.619628 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xw2k5\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-kube-api-access-xw2k5\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.643633 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-server-conf" (OuterVolumeSpecName: "server-conf") pod "82c8133d-d299-4c99-b8a0-96099e834e8c" (UID: "82c8133d-d299-4c99-b8a0-96099e834e8c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.646977 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.660556 4872 generic.go:334] "Generic (PLEG): container finished" podID="82c8133d-d299-4c99-b8a0-96099e834e8c" containerID="6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a" exitCode=0 Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.660617 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"82c8133d-d299-4c99-b8a0-96099e834e8c","Type":"ContainerDied","Data":"6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a"} Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.660646 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"82c8133d-d299-4c99-b8a0-96099e834e8c","Type":"ContainerDied","Data":"591c59aa46056780153b6968955e58429d6684e7d18c90693cde199497de5c5f"} Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.660664 4872 scope.go:117] "RemoveContainer" containerID="6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.660783 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.666328 4872 generic.go:334] "Generic (PLEG): container finished" podID="fca44d96-a000-4bf2-8283-a937b0192880" containerID="806f909f0f07687dd6ea8e62ed9234f07922743fa23a5604fb5f62e36f63f8b0" exitCode=0 Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.666356 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerDied","Data":"806f909f0f07687dd6ea8e62ed9234f07922743fa23a5604fb5f62e36f63f8b0"} Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.684120 4872 scope.go:117] "RemoveContainer" containerID="ad7fd4226f3c6d842b398d7962b8022bc46582f511d8ebd06feb9754c684ed0d" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.705683 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "82c8133d-d299-4c99-b8a0-96099e834e8c" (UID: "82c8133d-d299-4c99-b8a0-96099e834e8c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.721086 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82c8133d-d299-4c99-b8a0-96099e834e8c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.721122 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.721134 4872 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82c8133d-d299-4c99-b8a0-96099e834e8c-server-conf\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.723427 4872 scope.go:117] "RemoveContainer" containerID="6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a" Jan 26 09:27:40 crc kubenswrapper[4872]: E0126 09:27:40.723971 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a\": container with ID starting with 6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a not found: ID does not exist" containerID="6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.724023 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a"} err="failed to get container status \"6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a\": rpc error: code = NotFound desc = could not find container \"6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a\": container with ID starting with 6c9bd01f12bab2fdbbb4fac593a3b7dc1182f79e1ab9c9b11a0b656f89fde22a not found: ID does not exist" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.724053 4872 scope.go:117] "RemoveContainer" containerID="ad7fd4226f3c6d842b398d7962b8022bc46582f511d8ebd06feb9754c684ed0d" Jan 26 09:27:40 crc kubenswrapper[4872]: E0126 09:27:40.724474 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad7fd4226f3c6d842b398d7962b8022bc46582f511d8ebd06feb9754c684ed0d\": container with ID starting with ad7fd4226f3c6d842b398d7962b8022bc46582f511d8ebd06feb9754c684ed0d not found: ID does not exist" containerID="ad7fd4226f3c6d842b398d7962b8022bc46582f511d8ebd06feb9754c684ed0d" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.724528 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad7fd4226f3c6d842b398d7962b8022bc46582f511d8ebd06feb9754c684ed0d"} err="failed to get container status \"ad7fd4226f3c6d842b398d7962b8022bc46582f511d8ebd06feb9754c684ed0d\": rpc error: code = NotFound desc = could not find container \"ad7fd4226f3c6d842b398d7962b8022bc46582f511d8ebd06feb9754c684ed0d\": container with ID starting with ad7fd4226f3c6d842b398d7962b8022bc46582f511d8ebd06feb9754c684ed0d not found: ID does not exist" Jan 26 09:27:40 crc kubenswrapper[4872]: I0126 09:27:40.724567 4872 scope.go:117] "RemoveContainer" containerID="ad422c6c34b0fdde5ba287cc03a049a9b7acc36305d83e2c495ecd1a880a0a0f" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.114490 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.130638 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.159534 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 09:27:41 crc kubenswrapper[4872]: E0126 09:27:41.160090 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c8133d-d299-4c99-b8a0-96099e834e8c" containerName="setup-container" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.160106 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c8133d-d299-4c99-b8a0-96099e834e8c" containerName="setup-container" Jan 26 09:27:41 crc kubenswrapper[4872]: E0126 09:27:41.160117 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c8133d-d299-4c99-b8a0-96099e834e8c" containerName="rabbitmq" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.160123 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c8133d-d299-4c99-b8a0-96099e834e8c" containerName="rabbitmq" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.160302 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c8133d-d299-4c99-b8a0-96099e834e8c" containerName="rabbitmq" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.164034 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.169099 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.169351 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.169687 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.169861 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.170363 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.170398 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-5sngq" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.170555 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.171558 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.206727 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82c8133d-d299-4c99-b8a0-96099e834e8c" path="/var/lib/kubelet/pods/82c8133d-d299-4c99-b8a0-96099e834e8c/volumes" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.299749 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.344910 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fac8c53c-96c3-491f-9555-16a797c45699-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.345217 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fac8c53c-96c3-491f-9555-16a797c45699-pod-info\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.345430 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fac8c53c-96c3-491f-9555-16a797c45699-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.345513 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.345822 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fac8c53c-96c3-491f-9555-16a797c45699-server-conf\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.345867 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fac8c53c-96c3-491f-9555-16a797c45699-config-data\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.345929 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fac8c53c-96c3-491f-9555-16a797c45699-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.346004 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fac8c53c-96c3-491f-9555-16a797c45699-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.346088 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fac8c53c-96c3-491f-9555-16a797c45699-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.346119 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fac8c53c-96c3-491f-9555-16a797c45699-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.346623 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lhkl\" (UniqueName: \"kubernetes.io/projected/fac8c53c-96c3-491f-9555-16a797c45699-kube-api-access-8lhkl\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.449523 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-erlang-cookie-secret\") pod \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.449722 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-erlang-cookie\") pod \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.449750 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-server-conf\") pod \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.449845 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-pod-info\") pod \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.449893 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.449921 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-plugins\") pod \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.449955 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-config-data\") pod \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450028 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-plugins-conf\") pod \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450087 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-tls\") pod \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450132 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc9lx\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-kube-api-access-xc9lx\") pod \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450169 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-confd\") pod \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\" (UID: \"0792ad43-38af-4c94-bd3b-fc7d3a7af5da\") " Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450472 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fac8c53c-96c3-491f-9555-16a797c45699-server-conf\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450512 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fac8c53c-96c3-491f-9555-16a797c45699-config-data\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450547 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fac8c53c-96c3-491f-9555-16a797c45699-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450581 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fac8c53c-96c3-491f-9555-16a797c45699-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450611 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fac8c53c-96c3-491f-9555-16a797c45699-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450631 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fac8c53c-96c3-491f-9555-16a797c45699-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450687 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lhkl\" (UniqueName: \"kubernetes.io/projected/fac8c53c-96c3-491f-9555-16a797c45699-kube-api-access-8lhkl\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450744 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fac8c53c-96c3-491f-9555-16a797c45699-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.450838 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fac8c53c-96c3-491f-9555-16a797c45699-pod-info\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.451500 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fac8c53c-96c3-491f-9555-16a797c45699-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.451535 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.451877 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.452263 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fac8c53c-96c3-491f-9555-16a797c45699-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.453771 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fac8c53c-96c3-491f-9555-16a797c45699-server-conf\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.454547 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fac8c53c-96c3-491f-9555-16a797c45699-config-data\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.467036 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fac8c53c-96c3-491f-9555-16a797c45699-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.468171 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fac8c53c-96c3-491f-9555-16a797c45699-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.473396 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0792ad43-38af-4c94-bd3b-fc7d3a7af5da" (UID: "0792ad43-38af-4c94-bd3b-fc7d3a7af5da"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.478412 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fac8c53c-96c3-491f-9555-16a797c45699-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.479848 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0792ad43-38af-4c94-bd3b-fc7d3a7af5da" (UID: "0792ad43-38af-4c94-bd3b-fc7d3a7af5da"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.481783 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0792ad43-38af-4c94-bd3b-fc7d3a7af5da" (UID: "0792ad43-38af-4c94-bd3b-fc7d3a7af5da"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.508140 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-kube-api-access-xc9lx" (OuterVolumeSpecName: "kube-api-access-xc9lx") pod "0792ad43-38af-4c94-bd3b-fc7d3a7af5da" (UID: "0792ad43-38af-4c94-bd3b-fc7d3a7af5da"). InnerVolumeSpecName "kube-api-access-xc9lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.508446 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fac8c53c-96c3-491f-9555-16a797c45699-pod-info\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.509306 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fac8c53c-96c3-491f-9555-16a797c45699-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.510161 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-pod-info" (OuterVolumeSpecName: "pod-info") pod "0792ad43-38af-4c94-bd3b-fc7d3a7af5da" (UID: "0792ad43-38af-4c94-bd3b-fc7d3a7af5da"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.510606 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0792ad43-38af-4c94-bd3b-fc7d3a7af5da" (UID: "0792ad43-38af-4c94-bd3b-fc7d3a7af5da"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.516035 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "0792ad43-38af-4c94-bd3b-fc7d3a7af5da" (UID: "0792ad43-38af-4c94-bd3b-fc7d3a7af5da"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.522158 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fac8c53c-96c3-491f-9555-16a797c45699-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.537972 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0792ad43-38af-4c94-bd3b-fc7d3a7af5da" (UID: "0792ad43-38af-4c94-bd3b-fc7d3a7af5da"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.547669 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lhkl\" (UniqueName: \"kubernetes.io/projected/fac8c53c-96c3-491f-9555-16a797c45699-kube-api-access-8lhkl\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.558091 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.558124 4872 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-pod-info\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.558149 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.558158 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.558169 4872 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-plugins-conf\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.558176 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.558186 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc9lx\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-kube-api-access-xc9lx\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.558194 4872 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.598151 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-config-data" (OuterVolumeSpecName: "config-data") pod "0792ad43-38af-4c94-bd3b-fc7d3a7af5da" (UID: "0792ad43-38af-4c94-bd3b-fc7d3a7af5da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.615690 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.649839 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"fac8c53c-96c3-491f-9555-16a797c45699\") " pod="openstack/rabbitmq-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.660162 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.660202 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-config-data\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.682232 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerStarted","Data":"16869c8c2bd19091ae76a4591e6a82aaa7c0510c4b2bb4b199056a62beb5e880"} Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.684669 4872 generic.go:334] "Generic (PLEG): container finished" podID="0792ad43-38af-4c94-bd3b-fc7d3a7af5da" containerID="2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5" exitCode=0 Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.684740 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0792ad43-38af-4c94-bd3b-fc7d3a7af5da","Type":"ContainerDied","Data":"2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5"} Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.684767 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0792ad43-38af-4c94-bd3b-fc7d3a7af5da","Type":"ContainerDied","Data":"5f771a049c7d1063e46471249c3a8b5782df08fa31a10153c65849443943a9a4"} Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.684786 4872 scope.go:117] "RemoveContainer" containerID="2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.684744 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.688940 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-server-conf" (OuterVolumeSpecName: "server-conf") pod "0792ad43-38af-4c94-bd3b-fc7d3a7af5da" (UID: "0792ad43-38af-4c94-bd3b-fc7d3a7af5da"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.709705 4872 scope.go:117] "RemoveContainer" containerID="61e80e60ff53238778a51ec238b504652ca6a6c35b0dc148831642ff362f238a" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.735159 4872 scope.go:117] "RemoveContainer" containerID="2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5" Jan 26 09:27:41 crc kubenswrapper[4872]: E0126 09:27:41.735825 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5\": container with ID starting with 2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5 not found: ID does not exist" containerID="2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.735878 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5"} err="failed to get container status \"2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5\": rpc error: code = NotFound desc = could not find container \"2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5\": container with ID starting with 2f83bbc4eb6c1d10ca3d3def2a0cdf06e8e21815e9df666c1941fbe830d449b5 not found: ID does not exist" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.735915 4872 scope.go:117] "RemoveContainer" containerID="61e80e60ff53238778a51ec238b504652ca6a6c35b0dc148831642ff362f238a" Jan 26 09:27:41 crc kubenswrapper[4872]: E0126 09:27:41.736309 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61e80e60ff53238778a51ec238b504652ca6a6c35b0dc148831642ff362f238a\": container with ID starting with 61e80e60ff53238778a51ec238b504652ca6a6c35b0dc148831642ff362f238a not found: ID does not exist" containerID="61e80e60ff53238778a51ec238b504652ca6a6c35b0dc148831642ff362f238a" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.736355 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e80e60ff53238778a51ec238b504652ca6a6c35b0dc148831642ff362f238a"} err="failed to get container status \"61e80e60ff53238778a51ec238b504652ca6a6c35b0dc148831642ff362f238a\": rpc error: code = NotFound desc = could not find container \"61e80e60ff53238778a51ec238b504652ca6a6c35b0dc148831642ff362f238a\": container with ID starting with 61e80e60ff53238778a51ec238b504652ca6a6c35b0dc148831642ff362f238a not found: ID does not exist" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.742634 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0792ad43-38af-4c94-bd3b-fc7d3a7af5da" (UID: "0792ad43-38af-4c94-bd3b-fc7d3a7af5da"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.762951 4872 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-server-conf\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.763005 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0792ad43-38af-4c94-bd3b-fc7d3a7af5da-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Jan 26 09:27:41 crc kubenswrapper[4872]: I0126 09:27:41.784965 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.092732 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.104700 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.121570 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.151343 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 09:27:42 crc kubenswrapper[4872]: E0126 09:27:42.153299 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0792ad43-38af-4c94-bd3b-fc7d3a7af5da" containerName="rabbitmq" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.153332 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0792ad43-38af-4c94-bd3b-fc7d3a7af5da" containerName="rabbitmq" Jan 26 09:27:42 crc kubenswrapper[4872]: E0126 09:27:42.153353 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0792ad43-38af-4c94-bd3b-fc7d3a7af5da" containerName="setup-container" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.153360 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0792ad43-38af-4c94-bd3b-fc7d3a7af5da" containerName="setup-container" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.153565 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="0792ad43-38af-4c94-bd3b-fc7d3a7af5da" containerName="rabbitmq" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.154616 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.157024 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.158173 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.158480 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.158537 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.158608 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-grxrk" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.158831 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.159079 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.174380 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.277988 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3900451f-25bd-4b04-b516-20ff2a5c8233-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.278086 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7trnq\" (UniqueName: \"kubernetes.io/projected/3900451f-25bd-4b04-b516-20ff2a5c8233-kube-api-access-7trnq\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.278124 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.278149 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3900451f-25bd-4b04-b516-20ff2a5c8233-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.278168 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3900451f-25bd-4b04-b516-20ff2a5c8233-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.278253 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3900451f-25bd-4b04-b516-20ff2a5c8233-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.278357 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3900451f-25bd-4b04-b516-20ff2a5c8233-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.278999 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3900451f-25bd-4b04-b516-20ff2a5c8233-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.279074 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3900451f-25bd-4b04-b516-20ff2a5c8233-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.279169 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3900451f-25bd-4b04-b516-20ff2a5c8233-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.279246 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3900451f-25bd-4b04-b516-20ff2a5c8233-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.381324 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3900451f-25bd-4b04-b516-20ff2a5c8233-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.381402 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7trnq\" (UniqueName: \"kubernetes.io/projected/3900451f-25bd-4b04-b516-20ff2a5c8233-kube-api-access-7trnq\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.381475 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.381493 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3900451f-25bd-4b04-b516-20ff2a5c8233-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.381511 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3900451f-25bd-4b04-b516-20ff2a5c8233-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.381562 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3900451f-25bd-4b04-b516-20ff2a5c8233-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.381597 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3900451f-25bd-4b04-b516-20ff2a5c8233-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.381665 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3900451f-25bd-4b04-b516-20ff2a5c8233-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.382007 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3900451f-25bd-4b04-b516-20ff2a5c8233-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.382035 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3900451f-25bd-4b04-b516-20ff2a5c8233-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.382080 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3900451f-25bd-4b04-b516-20ff2a5c8233-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.382139 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3900451f-25bd-4b04-b516-20ff2a5c8233-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.381910 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.383277 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3900451f-25bd-4b04-b516-20ff2a5c8233-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.383470 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3900451f-25bd-4b04-b516-20ff2a5c8233-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.383624 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3900451f-25bd-4b04-b516-20ff2a5c8233-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.383739 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3900451f-25bd-4b04-b516-20ff2a5c8233-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.387276 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3900451f-25bd-4b04-b516-20ff2a5c8233-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.387314 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3900451f-25bd-4b04-b516-20ff2a5c8233-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.387642 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3900451f-25bd-4b04-b516-20ff2a5c8233-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.387717 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3900451f-25bd-4b04-b516-20ff2a5c8233-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.400447 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7trnq\" (UniqueName: \"kubernetes.io/projected/3900451f-25bd-4b04-b516-20ff2a5c8233-kube-api-access-7trnq\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.410316 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3900451f-25bd-4b04-b516-20ff2a5c8233\") " pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.579641 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:27:42 crc kubenswrapper[4872]: I0126 09:27:42.705294 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fac8c53c-96c3-491f-9555-16a797c45699","Type":"ContainerStarted","Data":"70c9e8a9be55183528e989774d584a5d3af3933f3ce0d254c847435c4b45c596"} Jan 26 09:27:43 crc kubenswrapper[4872]: I0126 09:27:43.131155 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Jan 26 09:27:43 crc kubenswrapper[4872]: I0126 09:27:43.196066 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0792ad43-38af-4c94-bd3b-fc7d3a7af5da" path="/var/lib/kubelet/pods/0792ad43-38af-4c94-bd3b-fc7d3a7af5da/volumes" Jan 26 09:27:43 crc kubenswrapper[4872]: I0126 09:27:43.715651 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3900451f-25bd-4b04-b516-20ff2a5c8233","Type":"ContainerStarted","Data":"8eb67c39305b14e6b0eafceb97860eb21e86d24fb3783926cee93a7640ed2d38"} Jan 26 09:27:44 crc kubenswrapper[4872]: I0126 09:27:44.735050 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fac8c53c-96c3-491f-9555-16a797c45699","Type":"ContainerStarted","Data":"c16e4f315b0eff42bc240993f178b1307bc610a5a7a2ff9095fc13d549a9caaa"} Jan 26 09:27:45 crc kubenswrapper[4872]: I0126 09:27:45.747920 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3900451f-25bd-4b04-b516-20ff2a5c8233","Type":"ContainerStarted","Data":"43710856f5852eb8f72d492609dd7cc6aa6fb0e08218c51709e07b33c00f67f6"} Jan 26 09:28:17 crc kubenswrapper[4872]: I0126 09:28:17.120323 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac8c53c-96c3-491f-9555-16a797c45699" containerID="c16e4f315b0eff42bc240993f178b1307bc610a5a7a2ff9095fc13d549a9caaa" exitCode=0 Jan 26 09:28:17 crc kubenswrapper[4872]: I0126 09:28:17.120442 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fac8c53c-96c3-491f-9555-16a797c45699","Type":"ContainerDied","Data":"c16e4f315b0eff42bc240993f178b1307bc610a5a7a2ff9095fc13d549a9caaa"} Jan 26 09:28:18 crc kubenswrapper[4872]: I0126 09:28:18.136195 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fac8c53c-96c3-491f-9555-16a797c45699","Type":"ContainerStarted","Data":"2b985baae3061dad47c3db936e214f1934c55975e94c7e05591de6875431c9e7"} Jan 26 09:28:18 crc kubenswrapper[4872]: I0126 09:28:18.136819 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Jan 26 09:28:18 crc kubenswrapper[4872]: I0126 09:28:18.140992 4872 generic.go:334] "Generic (PLEG): container finished" podID="3900451f-25bd-4b04-b516-20ff2a5c8233" containerID="43710856f5852eb8f72d492609dd7cc6aa6fb0e08218c51709e07b33c00f67f6" exitCode=0 Jan 26 09:28:18 crc kubenswrapper[4872]: I0126 09:28:18.141051 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3900451f-25bd-4b04-b516-20ff2a5c8233","Type":"ContainerDied","Data":"43710856f5852eb8f72d492609dd7cc6aa6fb0e08218c51709e07b33c00f67f6"} Jan 26 09:28:18 crc kubenswrapper[4872]: I0126 09:28:18.167781 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.167763954 podStartE2EDuration="37.167763954s" podCreationTimestamp="2026-01-26 09:27:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:28:18.165224998 +0000 UTC m=+1231.474064809" watchObservedRunningTime="2026-01-26 09:28:18.167763954 +0000 UTC m=+1231.476603765" Jan 26 09:28:19 crc kubenswrapper[4872]: I0126 09:28:19.152777 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3900451f-25bd-4b04-b516-20ff2a5c8233","Type":"ContainerStarted","Data":"1f18cf2496200fcab0e396b71c0ebaf5c9d79e9ea67f2b96ac1a6f4cef2b2597"} Jan 26 09:28:19 crc kubenswrapper[4872]: I0126 09:28:19.153541 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:28:19 crc kubenswrapper[4872]: I0126 09:28:19.193918 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.193891015 podStartE2EDuration="37.193891015s" podCreationTimestamp="2026-01-26 09:27:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-01-26 09:28:19.192787707 +0000 UTC m=+1232.501627548" watchObservedRunningTime="2026-01-26 09:28:19.193891015 +0000 UTC m=+1232.502730856" Jan 26 09:28:31 crc kubenswrapper[4872]: I0126 09:28:31.789092 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Jan 26 09:28:32 crc kubenswrapper[4872]: I0126 09:28:32.582994 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Jan 26 09:29:03 crc kubenswrapper[4872]: I0126 09:29:03.959723 4872 scope.go:117] "RemoveContainer" containerID="cbd60b0ea6d59b8ed7de42a25443218b3a80555a9f30aec90412ae7d48fa9ab2" Jan 26 09:29:04 crc kubenswrapper[4872]: I0126 09:29:04.001633 4872 scope.go:117] "RemoveContainer" containerID="cca8635ce74eaf63ddace9a7113b3ad50e65c19034ad7968f82733272bbd826e" Jan 26 09:29:04 crc kubenswrapper[4872]: I0126 09:29:04.083429 4872 scope.go:117] "RemoveContainer" containerID="8469ae4b59baa03382c6cc7faa957ae29ebde8d9f27e2a58f0d07160e555b42a" Jan 26 09:29:40 crc kubenswrapper[4872]: I0126 09:29:40.228163 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:29:40 crc kubenswrapper[4872]: I0126 09:29:40.228892 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.157741 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5"] Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.159990 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.164089 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.164464 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.183508 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5"] Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.308599 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv9z9\" (UniqueName: \"kubernetes.io/projected/93295990-db92-4fa2-8404-9ff1d4937979-kube-api-access-dv9z9\") pod \"collect-profiles-29490330-pb4l5\" (UID: \"93295990-db92-4fa2-8404-9ff1d4937979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.308759 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93295990-db92-4fa2-8404-9ff1d4937979-config-volume\") pod \"collect-profiles-29490330-pb4l5\" (UID: \"93295990-db92-4fa2-8404-9ff1d4937979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.309057 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93295990-db92-4fa2-8404-9ff1d4937979-secret-volume\") pod \"collect-profiles-29490330-pb4l5\" (UID: \"93295990-db92-4fa2-8404-9ff1d4937979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.410977 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93295990-db92-4fa2-8404-9ff1d4937979-config-volume\") pod \"collect-profiles-29490330-pb4l5\" (UID: \"93295990-db92-4fa2-8404-9ff1d4937979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.411143 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93295990-db92-4fa2-8404-9ff1d4937979-secret-volume\") pod \"collect-profiles-29490330-pb4l5\" (UID: \"93295990-db92-4fa2-8404-9ff1d4937979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.411178 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv9z9\" (UniqueName: \"kubernetes.io/projected/93295990-db92-4fa2-8404-9ff1d4937979-kube-api-access-dv9z9\") pod \"collect-profiles-29490330-pb4l5\" (UID: \"93295990-db92-4fa2-8404-9ff1d4937979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.412883 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93295990-db92-4fa2-8404-9ff1d4937979-config-volume\") pod \"collect-profiles-29490330-pb4l5\" (UID: \"93295990-db92-4fa2-8404-9ff1d4937979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.419663 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93295990-db92-4fa2-8404-9ff1d4937979-secret-volume\") pod \"collect-profiles-29490330-pb4l5\" (UID: \"93295990-db92-4fa2-8404-9ff1d4937979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.438562 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv9z9\" (UniqueName: \"kubernetes.io/projected/93295990-db92-4fa2-8404-9ff1d4937979-kube-api-access-dv9z9\") pod \"collect-profiles-29490330-pb4l5\" (UID: \"93295990-db92-4fa2-8404-9ff1d4937979\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.488700 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:00 crc kubenswrapper[4872]: I0126 09:30:00.937405 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5"] Jan 26 09:30:01 crc kubenswrapper[4872]: I0126 09:30:01.347528 4872 generic.go:334] "Generic (PLEG): container finished" podID="93295990-db92-4fa2-8404-9ff1d4937979" containerID="b8ecf69f7cbbe5907b3771eca3e60dbbee56b1494fc5446a8053bba8adec4d7b" exitCode=0 Jan 26 09:30:01 crc kubenswrapper[4872]: I0126 09:30:01.347587 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" event={"ID":"93295990-db92-4fa2-8404-9ff1d4937979","Type":"ContainerDied","Data":"b8ecf69f7cbbe5907b3771eca3e60dbbee56b1494fc5446a8053bba8adec4d7b"} Jan 26 09:30:01 crc kubenswrapper[4872]: I0126 09:30:01.347840 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" event={"ID":"93295990-db92-4fa2-8404-9ff1d4937979","Type":"ContainerStarted","Data":"0535ca69e214a51c06bc4551809611e61f8480eb5af8c5b27a79314210d2a7ff"} Jan 26 09:30:02 crc kubenswrapper[4872]: I0126 09:30:02.733221 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:02 crc kubenswrapper[4872]: I0126 09:30:02.862613 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dv9z9\" (UniqueName: \"kubernetes.io/projected/93295990-db92-4fa2-8404-9ff1d4937979-kube-api-access-dv9z9\") pod \"93295990-db92-4fa2-8404-9ff1d4937979\" (UID: \"93295990-db92-4fa2-8404-9ff1d4937979\") " Jan 26 09:30:02 crc kubenswrapper[4872]: I0126 09:30:02.863005 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93295990-db92-4fa2-8404-9ff1d4937979-secret-volume\") pod \"93295990-db92-4fa2-8404-9ff1d4937979\" (UID: \"93295990-db92-4fa2-8404-9ff1d4937979\") " Jan 26 09:30:02 crc kubenswrapper[4872]: I0126 09:30:02.863145 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93295990-db92-4fa2-8404-9ff1d4937979-config-volume\") pod \"93295990-db92-4fa2-8404-9ff1d4937979\" (UID: \"93295990-db92-4fa2-8404-9ff1d4937979\") " Jan 26 09:30:02 crc kubenswrapper[4872]: I0126 09:30:02.878100 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93295990-db92-4fa2-8404-9ff1d4937979-config-volume" (OuterVolumeSpecName: "config-volume") pod "93295990-db92-4fa2-8404-9ff1d4937979" (UID: "93295990-db92-4fa2-8404-9ff1d4937979"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:30:02 crc kubenswrapper[4872]: I0126 09:30:02.883230 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93295990-db92-4fa2-8404-9ff1d4937979-kube-api-access-dv9z9" (OuterVolumeSpecName: "kube-api-access-dv9z9") pod "93295990-db92-4fa2-8404-9ff1d4937979" (UID: "93295990-db92-4fa2-8404-9ff1d4937979"). InnerVolumeSpecName "kube-api-access-dv9z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:30:02 crc kubenswrapper[4872]: I0126 09:30:02.883842 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93295990-db92-4fa2-8404-9ff1d4937979-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "93295990-db92-4fa2-8404-9ff1d4937979" (UID: "93295990-db92-4fa2-8404-9ff1d4937979"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:30:02 crc kubenswrapper[4872]: I0126 09:30:02.965868 4872 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/93295990-db92-4fa2-8404-9ff1d4937979-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 09:30:02 crc kubenswrapper[4872]: I0126 09:30:02.966137 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dv9z9\" (UniqueName: \"kubernetes.io/projected/93295990-db92-4fa2-8404-9ff1d4937979-kube-api-access-dv9z9\") on node \"crc\" DevicePath \"\"" Jan 26 09:30:02 crc kubenswrapper[4872]: I0126 09:30:02.966148 4872 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/93295990-db92-4fa2-8404-9ff1d4937979-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 09:30:03 crc kubenswrapper[4872]: I0126 09:30:03.380918 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" event={"ID":"93295990-db92-4fa2-8404-9ff1d4937979","Type":"ContainerDied","Data":"0535ca69e214a51c06bc4551809611e61f8480eb5af8c5b27a79314210d2a7ff"} Jan 26 09:30:03 crc kubenswrapper[4872]: I0126 09:30:03.380966 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0535ca69e214a51c06bc4551809611e61f8480eb5af8c5b27a79314210d2a7ff" Jan 26 09:30:03 crc kubenswrapper[4872]: I0126 09:30:03.381014 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490330-pb4l5" Jan 26 09:30:03 crc kubenswrapper[4872]: E0126 09:30:03.423017 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93295990_db92_4fa2_8404_9ff1d4937979.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93295990_db92_4fa2_8404_9ff1d4937979.slice/crio-0535ca69e214a51c06bc4551809611e61f8480eb5af8c5b27a79314210d2a7ff\": RecentStats: unable to find data in memory cache]" Jan 26 09:30:04 crc kubenswrapper[4872]: I0126 09:30:04.223226 4872 scope.go:117] "RemoveContainer" containerID="d44c7c2ca90b6cfdc141b88d0cec514236df22703ee1b74ced2429b2657294ef" Jan 26 09:30:10 crc kubenswrapper[4872]: I0126 09:30:10.227466 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:30:10 crc kubenswrapper[4872]: I0126 09:30:10.228151 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.220194 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w6nr2"] Jan 26 09:30:15 crc kubenswrapper[4872]: E0126 09:30:15.221483 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93295990-db92-4fa2-8404-9ff1d4937979" containerName="collect-profiles" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.221507 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="93295990-db92-4fa2-8404-9ff1d4937979" containerName="collect-profiles" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.221909 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="93295990-db92-4fa2-8404-9ff1d4937979" containerName="collect-profiles" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.224285 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.234301 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w6nr2"] Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.334678 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3895a12f-36c6-40cd-a1ac-be382eab9bd3-catalog-content\") pod \"community-operators-w6nr2\" (UID: \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\") " pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.334769 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj8rk\" (UniqueName: \"kubernetes.io/projected/3895a12f-36c6-40cd-a1ac-be382eab9bd3-kube-api-access-vj8rk\") pod \"community-operators-w6nr2\" (UID: \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\") " pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.335161 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3895a12f-36c6-40cd-a1ac-be382eab9bd3-utilities\") pod \"community-operators-w6nr2\" (UID: \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\") " pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.437827 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3895a12f-36c6-40cd-a1ac-be382eab9bd3-utilities\") pod \"community-operators-w6nr2\" (UID: \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\") " pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.438034 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3895a12f-36c6-40cd-a1ac-be382eab9bd3-catalog-content\") pod \"community-operators-w6nr2\" (UID: \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\") " pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.438086 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj8rk\" (UniqueName: \"kubernetes.io/projected/3895a12f-36c6-40cd-a1ac-be382eab9bd3-kube-api-access-vj8rk\") pod \"community-operators-w6nr2\" (UID: \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\") " pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.438495 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3895a12f-36c6-40cd-a1ac-be382eab9bd3-utilities\") pod \"community-operators-w6nr2\" (UID: \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\") " pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.438646 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3895a12f-36c6-40cd-a1ac-be382eab9bd3-catalog-content\") pod \"community-operators-w6nr2\" (UID: \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\") " pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.469706 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj8rk\" (UniqueName: \"kubernetes.io/projected/3895a12f-36c6-40cd-a1ac-be382eab9bd3-kube-api-access-vj8rk\") pod \"community-operators-w6nr2\" (UID: \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\") " pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.595370 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:30:15 crc kubenswrapper[4872]: I0126 09:30:15.935247 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w6nr2"] Jan 26 09:30:16 crc kubenswrapper[4872]: I0126 09:30:16.525031 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nr2" event={"ID":"3895a12f-36c6-40cd-a1ac-be382eab9bd3","Type":"ContainerStarted","Data":"598cd1664f364b7f4d7aef2bc802adfe64eb61c09ca4f32e556839e6d9f10c85"} Jan 26 09:30:26 crc kubenswrapper[4872]: I0126 09:30:26.708077 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="6faf686b-43c3-44f2-9e84-a142ce8192d1" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.175:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 09:30:28 crc kubenswrapper[4872]: I0126 09:30:28.767173 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ironic-conductor-0" podUID="e135958c-61c2-4199-94b3-2a8e6623310c" containerName="ironic-conductor" probeResult="failure" output="command timed out" Jan 26 09:30:31 crc kubenswrapper[4872]: I0126 09:30:31.750139 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="6faf686b-43c3-44f2-9e84-a142ce8192d1" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.175:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 09:30:36 crc kubenswrapper[4872]: I0126 09:30:36.792100 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="6faf686b-43c3-44f2-9e84-a142ce8192d1" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.175:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Jan 26 09:30:36 crc kubenswrapper[4872]: I0126 09:30:36.792919 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 26 09:30:36 crc kubenswrapper[4872]: I0126 09:30:36.793908 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cinder-scheduler" containerStatusID={"Type":"cri-o","ID":"fc880a061465bda9031475fc46e469f58b56bf2f94b8b30a0515fe76465f20ee"} pod="openstack/cinder-scheduler-0" containerMessage="Container cinder-scheduler failed liveness probe, will be restarted" Jan 26 09:30:36 crc kubenswrapper[4872]: I0126 09:30:36.793971 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6faf686b-43c3-44f2-9e84-a142ce8192d1" containerName="cinder-scheduler" containerID="cri-o://fc880a061465bda9031475fc46e469f58b56bf2f94b8b30a0515fe76465f20ee" gracePeriod=30 Jan 26 09:30:40 crc kubenswrapper[4872]: I0126 09:30:40.228159 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:30:40 crc kubenswrapper[4872]: I0126 09:30:40.229119 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:30:40 crc kubenswrapper[4872]: I0126 09:30:40.229176 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:30:40 crc kubenswrapper[4872]: I0126 09:30:40.230136 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"16869c8c2bd19091ae76a4591e6a82aaa7c0510c4b2bb4b199056a62beb5e880"} pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 09:30:40 crc kubenswrapper[4872]: I0126 09:30:40.230202 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" containerID="cri-o://16869c8c2bd19091ae76a4591e6a82aaa7c0510c4b2bb4b199056a62beb5e880" gracePeriod=600 Jan 26 09:30:48 crc kubenswrapper[4872]: I0126 09:30:48.609295 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_machine-config-daemon-gt4gn_fca44d96-a000-4bf2-8283-a937b0192880/machine-config-daemon/6.log" Jan 26 09:30:48 crc kubenswrapper[4872]: I0126 09:30:48.612198 4872 generic.go:334] "Generic (PLEG): container finished" podID="fca44d96-a000-4bf2-8283-a937b0192880" containerID="16869c8c2bd19091ae76a4591e6a82aaa7c0510c4b2bb4b199056a62beb5e880" exitCode=-1 Jan 26 09:30:48 crc kubenswrapper[4872]: I0126 09:30:48.612287 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerDied","Data":"16869c8c2bd19091ae76a4591e6a82aaa7c0510c4b2bb4b199056a62beb5e880"} Jan 26 09:30:48 crc kubenswrapper[4872]: I0126 09:30:48.612443 4872 scope.go:117] "RemoveContainer" containerID="806f909f0f07687dd6ea8e62ed9234f07922743fa23a5604fb5f62e36f63f8b0" Jan 26 09:30:58 crc kubenswrapper[4872]: I0126 09:30:58.767717 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ironic-conductor-0" podUID="e135958c-61c2-4199-94b3-2a8e6623310c" containerName="ironic-conductor" probeResult="failure" output="command timed out" Jan 26 09:31:04 crc kubenswrapper[4872]: I0126 09:31:04.319717 4872 scope.go:117] "RemoveContainer" containerID="c30824a7e633a670b986168841da9f3c10f88d1edbce9977dababd548e174e1e" Jan 26 09:31:22 crc kubenswrapper[4872]: I0126 09:31:22.730593 4872 generic.go:334] "Generic (PLEG): container finished" podID="6faf686b-43c3-44f2-9e84-a142ce8192d1" containerID="fc880a061465bda9031475fc46e469f58b56bf2f94b8b30a0515fe76465f20ee" exitCode=-1 Jan 26 09:31:22 crc kubenswrapper[4872]: I0126 09:31:22.730752 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6faf686b-43c3-44f2-9e84-a142ce8192d1","Type":"ContainerDied","Data":"fc880a061465bda9031475fc46e469f58b56bf2f94b8b30a0515fe76465f20ee"} Jan 26 09:31:26 crc kubenswrapper[4872]: I0126 09:31:26.787498 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerStarted","Data":"fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a"} Jan 26 09:31:26 crc kubenswrapper[4872]: I0126 09:31:26.789566 4872 generic.go:334] "Generic (PLEG): container finished" podID="3895a12f-36c6-40cd-a1ac-be382eab9bd3" containerID="8a5b47570f80b355c7c677bc6bca3f18bcdc4c4fc900a81e7dbead6282e461d1" exitCode=0 Jan 26 09:31:26 crc kubenswrapper[4872]: I0126 09:31:26.789609 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nr2" event={"ID":"3895a12f-36c6-40cd-a1ac-be382eab9bd3","Type":"ContainerDied","Data":"8a5b47570f80b355c7c677bc6bca3f18bcdc4c4fc900a81e7dbead6282e461d1"} Jan 26 09:31:27 crc kubenswrapper[4872]: I0126 09:31:27.805200 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6faf686b-43c3-44f2-9e84-a142ce8192d1","Type":"ContainerStarted","Data":"37ddbc1063e7cc5d5f3ad81f26de8579e6e007e6f6d20912d74b74f9bba583fa"} Jan 26 09:31:28 crc kubenswrapper[4872]: I0126 09:31:28.822846 4872 generic.go:334] "Generic (PLEG): container finished" podID="3895a12f-36c6-40cd-a1ac-be382eab9bd3" containerID="d35c03fedce973bfc148e031175cb30cf196a4b07fdb974ee882659c2464a14c" exitCode=0 Jan 26 09:31:28 crc kubenswrapper[4872]: I0126 09:31:28.824737 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nr2" event={"ID":"3895a12f-36c6-40cd-a1ac-be382eab9bd3","Type":"ContainerDied","Data":"d35c03fedce973bfc148e031175cb30cf196a4b07fdb974ee882659c2464a14c"} Jan 26 09:31:29 crc kubenswrapper[4872]: I0126 09:31:29.666278 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Jan 26 09:31:29 crc kubenswrapper[4872]: I0126 09:31:29.837937 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nr2" event={"ID":"3895a12f-36c6-40cd-a1ac-be382eab9bd3","Type":"ContainerStarted","Data":"179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f"} Jan 26 09:31:29 crc kubenswrapper[4872]: I0126 09:31:29.860459 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w6nr2" podStartSLOduration=72.164323317 podStartE2EDuration="1m14.86044067s" podCreationTimestamp="2026-01-26 09:30:15 +0000 UTC" firstStartedPulling="2026-01-26 09:31:26.791428904 +0000 UTC m=+1420.100268705" lastFinishedPulling="2026-01-26 09:31:29.487546217 +0000 UTC m=+1422.796386058" observedRunningTime="2026-01-26 09:31:29.858197143 +0000 UTC m=+1423.167036944" watchObservedRunningTime="2026-01-26 09:31:29.86044067 +0000 UTC m=+1423.169280471" Jan 26 09:31:34 crc kubenswrapper[4872]: I0126 09:31:34.674767 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Jan 26 09:31:35 crc kubenswrapper[4872]: I0126 09:31:35.595986 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:31:35 crc kubenswrapper[4872]: I0126 09:31:35.596308 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:31:35 crc kubenswrapper[4872]: I0126 09:31:35.656784 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:31:35 crc kubenswrapper[4872]: I0126 09:31:35.994089 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:31:36 crc kubenswrapper[4872]: I0126 09:31:36.063567 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w6nr2"] Jan 26 09:31:37 crc kubenswrapper[4872]: I0126 09:31:37.959718 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w6nr2" podUID="3895a12f-36c6-40cd-a1ac-be382eab9bd3" containerName="registry-server" containerID="cri-o://179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f" gracePeriod=2 Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.481418 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.620575 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3895a12f-36c6-40cd-a1ac-be382eab9bd3-utilities" (OuterVolumeSpecName: "utilities") pod "3895a12f-36c6-40cd-a1ac-be382eab9bd3" (UID: "3895a12f-36c6-40cd-a1ac-be382eab9bd3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.619791 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3895a12f-36c6-40cd-a1ac-be382eab9bd3-utilities\") pod \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\" (UID: \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\") " Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.620881 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vj8rk\" (UniqueName: \"kubernetes.io/projected/3895a12f-36c6-40cd-a1ac-be382eab9bd3-kube-api-access-vj8rk\") pod \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\" (UID: \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\") " Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.620918 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3895a12f-36c6-40cd-a1ac-be382eab9bd3-catalog-content\") pod \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\" (UID: \"3895a12f-36c6-40cd-a1ac-be382eab9bd3\") " Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.622090 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3895a12f-36c6-40cd-a1ac-be382eab9bd3-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.627635 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3895a12f-36c6-40cd-a1ac-be382eab9bd3-kube-api-access-vj8rk" (OuterVolumeSpecName: "kube-api-access-vj8rk") pod "3895a12f-36c6-40cd-a1ac-be382eab9bd3" (UID: "3895a12f-36c6-40cd-a1ac-be382eab9bd3"). InnerVolumeSpecName "kube-api-access-vj8rk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.706653 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3895a12f-36c6-40cd-a1ac-be382eab9bd3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3895a12f-36c6-40cd-a1ac-be382eab9bd3" (UID: "3895a12f-36c6-40cd-a1ac-be382eab9bd3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.724569 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vj8rk\" (UniqueName: \"kubernetes.io/projected/3895a12f-36c6-40cd-a1ac-be382eab9bd3-kube-api-access-vj8rk\") on node \"crc\" DevicePath \"\"" Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.724630 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3895a12f-36c6-40cd-a1ac-be382eab9bd3-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.976057 4872 generic.go:334] "Generic (PLEG): container finished" podID="3895a12f-36c6-40cd-a1ac-be382eab9bd3" containerID="179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f" exitCode=0 Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.976118 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w6nr2" Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.976123 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nr2" event={"ID":"3895a12f-36c6-40cd-a1ac-be382eab9bd3","Type":"ContainerDied","Data":"179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f"} Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.976303 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w6nr2" event={"ID":"3895a12f-36c6-40cd-a1ac-be382eab9bd3","Type":"ContainerDied","Data":"598cd1664f364b7f4d7aef2bc802adfe64eb61c09ca4f32e556839e6d9f10c85"} Jan 26 09:31:38 crc kubenswrapper[4872]: I0126 09:31:38.976795 4872 scope.go:117] "RemoveContainer" containerID="179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f" Jan 26 09:31:39 crc kubenswrapper[4872]: I0126 09:31:39.022039 4872 scope.go:117] "RemoveContainer" containerID="d35c03fedce973bfc148e031175cb30cf196a4b07fdb974ee882659c2464a14c" Jan 26 09:31:39 crc kubenswrapper[4872]: I0126 09:31:39.042431 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w6nr2"] Jan 26 09:31:39 crc kubenswrapper[4872]: I0126 09:31:39.053514 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w6nr2"] Jan 26 09:31:39 crc kubenswrapper[4872]: I0126 09:31:39.057604 4872 scope.go:117] "RemoveContainer" containerID="8a5b47570f80b355c7c677bc6bca3f18bcdc4c4fc900a81e7dbead6282e461d1" Jan 26 09:31:39 crc kubenswrapper[4872]: I0126 09:31:39.114555 4872 scope.go:117] "RemoveContainer" containerID="179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f" Jan 26 09:31:39 crc kubenswrapper[4872]: E0126 09:31:39.115012 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f\": container with ID starting with 179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f not found: ID does not exist" containerID="179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f" Jan 26 09:31:39 crc kubenswrapper[4872]: I0126 09:31:39.115056 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f"} err="failed to get container status \"179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f\": rpc error: code = NotFound desc = could not find container \"179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f\": container with ID starting with 179562ce986520447ba4b54dfc7c4577bcc49a48dca2a3813985b8fc801c4d1f not found: ID does not exist" Jan 26 09:31:39 crc kubenswrapper[4872]: I0126 09:31:39.115084 4872 scope.go:117] "RemoveContainer" containerID="d35c03fedce973bfc148e031175cb30cf196a4b07fdb974ee882659c2464a14c" Jan 26 09:31:39 crc kubenswrapper[4872]: E0126 09:31:39.115510 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d35c03fedce973bfc148e031175cb30cf196a4b07fdb974ee882659c2464a14c\": container with ID starting with d35c03fedce973bfc148e031175cb30cf196a4b07fdb974ee882659c2464a14c not found: ID does not exist" containerID="d35c03fedce973bfc148e031175cb30cf196a4b07fdb974ee882659c2464a14c" Jan 26 09:31:39 crc kubenswrapper[4872]: I0126 09:31:39.115538 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d35c03fedce973bfc148e031175cb30cf196a4b07fdb974ee882659c2464a14c"} err="failed to get container status \"d35c03fedce973bfc148e031175cb30cf196a4b07fdb974ee882659c2464a14c\": rpc error: code = NotFound desc = could not find container \"d35c03fedce973bfc148e031175cb30cf196a4b07fdb974ee882659c2464a14c\": container with ID starting with d35c03fedce973bfc148e031175cb30cf196a4b07fdb974ee882659c2464a14c not found: ID does not exist" Jan 26 09:31:39 crc kubenswrapper[4872]: I0126 09:31:39.115560 4872 scope.go:117] "RemoveContainer" containerID="8a5b47570f80b355c7c677bc6bca3f18bcdc4c4fc900a81e7dbead6282e461d1" Jan 26 09:31:39 crc kubenswrapper[4872]: E0126 09:31:39.115913 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a5b47570f80b355c7c677bc6bca3f18bcdc4c4fc900a81e7dbead6282e461d1\": container with ID starting with 8a5b47570f80b355c7c677bc6bca3f18bcdc4c4fc900a81e7dbead6282e461d1 not found: ID does not exist" containerID="8a5b47570f80b355c7c677bc6bca3f18bcdc4c4fc900a81e7dbead6282e461d1" Jan 26 09:31:39 crc kubenswrapper[4872]: I0126 09:31:39.115934 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a5b47570f80b355c7c677bc6bca3f18bcdc4c4fc900a81e7dbead6282e461d1"} err="failed to get container status \"8a5b47570f80b355c7c677bc6bca3f18bcdc4c4fc900a81e7dbead6282e461d1\": rpc error: code = NotFound desc = could not find container \"8a5b47570f80b355c7c677bc6bca3f18bcdc4c4fc900a81e7dbead6282e461d1\": container with ID starting with 8a5b47570f80b355c7c677bc6bca3f18bcdc4c4fc900a81e7dbead6282e461d1 not found: ID does not exist" Jan 26 09:31:39 crc kubenswrapper[4872]: I0126 09:31:39.196050 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3895a12f-36c6-40cd-a1ac-be382eab9bd3" path="/var/lib/kubelet/pods/3895a12f-36c6-40cd-a1ac-be382eab9bd3/volumes" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.386975 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k25ls"] Jan 26 09:31:53 crc kubenswrapper[4872]: E0126 09:31:53.388592 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3895a12f-36c6-40cd-a1ac-be382eab9bd3" containerName="extract-utilities" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.388627 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3895a12f-36c6-40cd-a1ac-be382eab9bd3" containerName="extract-utilities" Jan 26 09:31:53 crc kubenswrapper[4872]: E0126 09:31:53.388700 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3895a12f-36c6-40cd-a1ac-be382eab9bd3" containerName="extract-content" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.388721 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3895a12f-36c6-40cd-a1ac-be382eab9bd3" containerName="extract-content" Jan 26 09:31:53 crc kubenswrapper[4872]: E0126 09:31:53.388757 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3895a12f-36c6-40cd-a1ac-be382eab9bd3" containerName="registry-server" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.388777 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3895a12f-36c6-40cd-a1ac-be382eab9bd3" containerName="registry-server" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.389310 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3895a12f-36c6-40cd-a1ac-be382eab9bd3" containerName="registry-server" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.393590 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.423373 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k25ls"] Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.449229 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-utilities\") pod \"redhat-marketplace-k25ls\" (UID: \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\") " pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.449347 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6w64\" (UniqueName: \"kubernetes.io/projected/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-kube-api-access-j6w64\") pod \"redhat-marketplace-k25ls\" (UID: \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\") " pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.449458 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-catalog-content\") pod \"redhat-marketplace-k25ls\" (UID: \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\") " pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.551408 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-utilities\") pod \"redhat-marketplace-k25ls\" (UID: \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\") " pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.551589 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6w64\" (UniqueName: \"kubernetes.io/projected/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-kube-api-access-j6w64\") pod \"redhat-marketplace-k25ls\" (UID: \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\") " pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.551736 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-catalog-content\") pod \"redhat-marketplace-k25ls\" (UID: \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\") " pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.552582 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-catalog-content\") pod \"redhat-marketplace-k25ls\" (UID: \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\") " pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.553064 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-utilities\") pod \"redhat-marketplace-k25ls\" (UID: \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\") " pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.580292 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6w64\" (UniqueName: \"kubernetes.io/projected/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-kube-api-access-j6w64\") pod \"redhat-marketplace-k25ls\" (UID: \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\") " pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:31:53 crc kubenswrapper[4872]: I0126 09:31:53.728821 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:31:54 crc kubenswrapper[4872]: I0126 09:31:54.254183 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k25ls"] Jan 26 09:31:55 crc kubenswrapper[4872]: I0126 09:31:55.154817 4872 generic.go:334] "Generic (PLEG): container finished" podID="9eb2f617-9a5a-483c-aff5-3838bc2c68ec" containerID="3998716e7c3a8bd7d98feae912d1f3535beb82477a5e165bdae7320a9e42d717" exitCode=0 Jan 26 09:31:55 crc kubenswrapper[4872]: I0126 09:31:55.154947 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k25ls" event={"ID":"9eb2f617-9a5a-483c-aff5-3838bc2c68ec","Type":"ContainerDied","Data":"3998716e7c3a8bd7d98feae912d1f3535beb82477a5e165bdae7320a9e42d717"} Jan 26 09:31:55 crc kubenswrapper[4872]: I0126 09:31:55.155210 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k25ls" event={"ID":"9eb2f617-9a5a-483c-aff5-3838bc2c68ec","Type":"ContainerStarted","Data":"46e322021b8afe9abd4fbf69f073c11c5c7d15fb3d040959064bab176622d64c"} Jan 26 09:31:55 crc kubenswrapper[4872]: I0126 09:31:55.157387 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 09:31:56 crc kubenswrapper[4872]: E0126 09:31:56.537430 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9eb2f617_9a5a_483c_aff5_3838bc2c68ec.slice/crio-conmon-851e636ef67d3ab45d5b3a68ba8b6b60d17b6509ae0ebcf3c6b93997e3c02241.scope\": RecentStats: unable to find data in memory cache]" Jan 26 09:31:57 crc kubenswrapper[4872]: I0126 09:31:57.174587 4872 generic.go:334] "Generic (PLEG): container finished" podID="9eb2f617-9a5a-483c-aff5-3838bc2c68ec" containerID="851e636ef67d3ab45d5b3a68ba8b6b60d17b6509ae0ebcf3c6b93997e3c02241" exitCode=0 Jan 26 09:31:57 crc kubenswrapper[4872]: I0126 09:31:57.174654 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k25ls" event={"ID":"9eb2f617-9a5a-483c-aff5-3838bc2c68ec","Type":"ContainerDied","Data":"851e636ef67d3ab45d5b3a68ba8b6b60d17b6509ae0ebcf3c6b93997e3c02241"} Jan 26 09:31:58 crc kubenswrapper[4872]: I0126 09:31:58.190602 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k25ls" event={"ID":"9eb2f617-9a5a-483c-aff5-3838bc2c68ec","Type":"ContainerStarted","Data":"cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c"} Jan 26 09:31:58 crc kubenswrapper[4872]: I0126 09:31:58.216107 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k25ls" podStartSLOduration=2.688149413 podStartE2EDuration="5.216090854s" podCreationTimestamp="2026-01-26 09:31:53 +0000 UTC" firstStartedPulling="2026-01-26 09:31:55.157002501 +0000 UTC m=+1448.465842312" lastFinishedPulling="2026-01-26 09:31:57.684943952 +0000 UTC m=+1450.993783753" observedRunningTime="2026-01-26 09:31:58.209562569 +0000 UTC m=+1451.518402390" watchObservedRunningTime="2026-01-26 09:31:58.216090854 +0000 UTC m=+1451.524930655" Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.150928 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fxc65"] Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.153447 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.166704 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/336dfe89-6359-4d43-956f-8e9c119cf45e-catalog-content\") pod \"certified-operators-fxc65\" (UID: \"336dfe89-6359-4d43-956f-8e9c119cf45e\") " pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.167029 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/336dfe89-6359-4d43-956f-8e9c119cf45e-utilities\") pod \"certified-operators-fxc65\" (UID: \"336dfe89-6359-4d43-956f-8e9c119cf45e\") " pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.167132 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqjjh\" (UniqueName: \"kubernetes.io/projected/336dfe89-6359-4d43-956f-8e9c119cf45e-kube-api-access-nqjjh\") pod \"certified-operators-fxc65\" (UID: \"336dfe89-6359-4d43-956f-8e9c119cf45e\") " pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.174104 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fxc65"] Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.269383 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/336dfe89-6359-4d43-956f-8e9c119cf45e-utilities\") pod \"certified-operators-fxc65\" (UID: \"336dfe89-6359-4d43-956f-8e9c119cf45e\") " pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.269473 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqjjh\" (UniqueName: \"kubernetes.io/projected/336dfe89-6359-4d43-956f-8e9c119cf45e-kube-api-access-nqjjh\") pod \"certified-operators-fxc65\" (UID: \"336dfe89-6359-4d43-956f-8e9c119cf45e\") " pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.269509 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/336dfe89-6359-4d43-956f-8e9c119cf45e-catalog-content\") pod \"certified-operators-fxc65\" (UID: \"336dfe89-6359-4d43-956f-8e9c119cf45e\") " pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.270626 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/336dfe89-6359-4d43-956f-8e9c119cf45e-utilities\") pod \"certified-operators-fxc65\" (UID: \"336dfe89-6359-4d43-956f-8e9c119cf45e\") " pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.271107 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/336dfe89-6359-4d43-956f-8e9c119cf45e-catalog-content\") pod \"certified-operators-fxc65\" (UID: \"336dfe89-6359-4d43-956f-8e9c119cf45e\") " pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.290492 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqjjh\" (UniqueName: \"kubernetes.io/projected/336dfe89-6359-4d43-956f-8e9c119cf45e-kube-api-access-nqjjh\") pod \"certified-operators-fxc65\" (UID: \"336dfe89-6359-4d43-956f-8e9c119cf45e\") " pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.473010 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:31:59 crc kubenswrapper[4872]: I0126 09:31:59.962779 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fxc65"] Jan 26 09:31:59 crc kubenswrapper[4872]: W0126 09:31:59.972140 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod336dfe89_6359_4d43_956f_8e9c119cf45e.slice/crio-0adcd2e5f5d6d11cec42895422771396f0c880b3dbc835150cbf6904c62a2886 WatchSource:0}: Error finding container 0adcd2e5f5d6d11cec42895422771396f0c880b3dbc835150cbf6904c62a2886: Status 404 returned error can't find the container with id 0adcd2e5f5d6d11cec42895422771396f0c880b3dbc835150cbf6904c62a2886 Jan 26 09:32:00 crc kubenswrapper[4872]: I0126 09:32:00.246427 4872 generic.go:334] "Generic (PLEG): container finished" podID="336dfe89-6359-4d43-956f-8e9c119cf45e" containerID="a95f31d902c0c641587bbf6f34859784acc03754639389862c0d97c82ab4124c" exitCode=0 Jan 26 09:32:00 crc kubenswrapper[4872]: I0126 09:32:00.246489 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxc65" event={"ID":"336dfe89-6359-4d43-956f-8e9c119cf45e","Type":"ContainerDied","Data":"a95f31d902c0c641587bbf6f34859784acc03754639389862c0d97c82ab4124c"} Jan 26 09:32:00 crc kubenswrapper[4872]: I0126 09:32:00.246958 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxc65" event={"ID":"336dfe89-6359-4d43-956f-8e9c119cf45e","Type":"ContainerStarted","Data":"0adcd2e5f5d6d11cec42895422771396f0c880b3dbc835150cbf6904c62a2886"} Jan 26 09:32:02 crc kubenswrapper[4872]: I0126 09:32:02.282135 4872 generic.go:334] "Generic (PLEG): container finished" podID="336dfe89-6359-4d43-956f-8e9c119cf45e" containerID="68f1e2f5f9d9f9d7567d2882cbab053f9b1ec0a874b5df1bfd626c320963e6e5" exitCode=0 Jan 26 09:32:02 crc kubenswrapper[4872]: I0126 09:32:02.282761 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxc65" event={"ID":"336dfe89-6359-4d43-956f-8e9c119cf45e","Type":"ContainerDied","Data":"68f1e2f5f9d9f9d7567d2882cbab053f9b1ec0a874b5df1bfd626c320963e6e5"} Jan 26 09:32:03 crc kubenswrapper[4872]: I0126 09:32:03.293189 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxc65" event={"ID":"336dfe89-6359-4d43-956f-8e9c119cf45e","Type":"ContainerStarted","Data":"6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409"} Jan 26 09:32:03 crc kubenswrapper[4872]: I0126 09:32:03.316150 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fxc65" podStartSLOduration=1.717707861 podStartE2EDuration="4.316122699s" podCreationTimestamp="2026-01-26 09:31:59 +0000 UTC" firstStartedPulling="2026-01-26 09:32:00.248648995 +0000 UTC m=+1453.557488796" lastFinishedPulling="2026-01-26 09:32:02.847063833 +0000 UTC m=+1456.155903634" observedRunningTime="2026-01-26 09:32:03.312782595 +0000 UTC m=+1456.621622446" watchObservedRunningTime="2026-01-26 09:32:03.316122699 +0000 UTC m=+1456.624962520" Jan 26 09:32:03 crc kubenswrapper[4872]: I0126 09:32:03.729097 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:32:03 crc kubenswrapper[4872]: I0126 09:32:03.729600 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:32:03 crc kubenswrapper[4872]: I0126 09:32:03.812750 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:32:04 crc kubenswrapper[4872]: I0126 09:32:04.355432 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:32:04 crc kubenswrapper[4872]: I0126 09:32:04.934908 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k25ls"] Jan 26 09:32:06 crc kubenswrapper[4872]: I0126 09:32:06.322336 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k25ls" podUID="9eb2f617-9a5a-483c-aff5-3838bc2c68ec" containerName="registry-server" containerID="cri-o://cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c" gracePeriod=2 Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.264107 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.332976 4872 generic.go:334] "Generic (PLEG): container finished" podID="9eb2f617-9a5a-483c-aff5-3838bc2c68ec" containerID="cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c" exitCode=0 Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.333017 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k25ls" event={"ID":"9eb2f617-9a5a-483c-aff5-3838bc2c68ec","Type":"ContainerDied","Data":"cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c"} Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.333045 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k25ls" event={"ID":"9eb2f617-9a5a-483c-aff5-3838bc2c68ec","Type":"ContainerDied","Data":"46e322021b8afe9abd4fbf69f073c11c5c7d15fb3d040959064bab176622d64c"} Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.333061 4872 scope.go:117] "RemoveContainer" containerID="cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.333963 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k25ls" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.356624 4872 scope.go:117] "RemoveContainer" containerID="851e636ef67d3ab45d5b3a68ba8b6b60d17b6509ae0ebcf3c6b93997e3c02241" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.385143 4872 scope.go:117] "RemoveContainer" containerID="3998716e7c3a8bd7d98feae912d1f3535beb82477a5e165bdae7320a9e42d717" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.426069 4872 scope.go:117] "RemoveContainer" containerID="cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c" Jan 26 09:32:07 crc kubenswrapper[4872]: E0126 09:32:07.427003 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c\": container with ID starting with cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c not found: ID does not exist" containerID="cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.427110 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c"} err="failed to get container status \"cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c\": rpc error: code = NotFound desc = could not find container \"cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c\": container with ID starting with cc4baa3d0ccfbe839357183e097c634c65225363aca88ac1ef5812d92bef1f0c not found: ID does not exist" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.427157 4872 scope.go:117] "RemoveContainer" containerID="851e636ef67d3ab45d5b3a68ba8b6b60d17b6509ae0ebcf3c6b93997e3c02241" Jan 26 09:32:07 crc kubenswrapper[4872]: E0126 09:32:07.427833 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"851e636ef67d3ab45d5b3a68ba8b6b60d17b6509ae0ebcf3c6b93997e3c02241\": container with ID starting with 851e636ef67d3ab45d5b3a68ba8b6b60d17b6509ae0ebcf3c6b93997e3c02241 not found: ID does not exist" containerID="851e636ef67d3ab45d5b3a68ba8b6b60d17b6509ae0ebcf3c6b93997e3c02241" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.427895 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"851e636ef67d3ab45d5b3a68ba8b6b60d17b6509ae0ebcf3c6b93997e3c02241"} err="failed to get container status \"851e636ef67d3ab45d5b3a68ba8b6b60d17b6509ae0ebcf3c6b93997e3c02241\": rpc error: code = NotFound desc = could not find container \"851e636ef67d3ab45d5b3a68ba8b6b60d17b6509ae0ebcf3c6b93997e3c02241\": container with ID starting with 851e636ef67d3ab45d5b3a68ba8b6b60d17b6509ae0ebcf3c6b93997e3c02241 not found: ID does not exist" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.427934 4872 scope.go:117] "RemoveContainer" containerID="3998716e7c3a8bd7d98feae912d1f3535beb82477a5e165bdae7320a9e42d717" Jan 26 09:32:07 crc kubenswrapper[4872]: E0126 09:32:07.428445 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3998716e7c3a8bd7d98feae912d1f3535beb82477a5e165bdae7320a9e42d717\": container with ID starting with 3998716e7c3a8bd7d98feae912d1f3535beb82477a5e165bdae7320a9e42d717 not found: ID does not exist" containerID="3998716e7c3a8bd7d98feae912d1f3535beb82477a5e165bdae7320a9e42d717" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.428673 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3998716e7c3a8bd7d98feae912d1f3535beb82477a5e165bdae7320a9e42d717"} err="failed to get container status \"3998716e7c3a8bd7d98feae912d1f3535beb82477a5e165bdae7320a9e42d717\": rpc error: code = NotFound desc = could not find container \"3998716e7c3a8bd7d98feae912d1f3535beb82477a5e165bdae7320a9e42d717\": container with ID starting with 3998716e7c3a8bd7d98feae912d1f3535beb82477a5e165bdae7320a9e42d717 not found: ID does not exist" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.467471 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-catalog-content\") pod \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\" (UID: \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\") " Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.467631 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6w64\" (UniqueName: \"kubernetes.io/projected/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-kube-api-access-j6w64\") pod \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\" (UID: \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\") " Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.467942 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-utilities\") pod \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\" (UID: \"9eb2f617-9a5a-483c-aff5-3838bc2c68ec\") " Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.468672 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-utilities" (OuterVolumeSpecName: "utilities") pod "9eb2f617-9a5a-483c-aff5-3838bc2c68ec" (UID: "9eb2f617-9a5a-483c-aff5-3838bc2c68ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.475359 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-kube-api-access-j6w64" (OuterVolumeSpecName: "kube-api-access-j6w64") pod "9eb2f617-9a5a-483c-aff5-3838bc2c68ec" (UID: "9eb2f617-9a5a-483c-aff5-3838bc2c68ec"). InnerVolumeSpecName "kube-api-access-j6w64". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.551147 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9eb2f617-9a5a-483c-aff5-3838bc2c68ec" (UID: "9eb2f617-9a5a-483c-aff5-3838bc2c68ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.572617 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.572648 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.572659 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6w64\" (UniqueName: \"kubernetes.io/projected/9eb2f617-9a5a-483c-aff5-3838bc2c68ec-kube-api-access-j6w64\") on node \"crc\" DevicePath \"\"" Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.669976 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k25ls"] Jan 26 09:32:07 crc kubenswrapper[4872]: I0126 09:32:07.679566 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k25ls"] Jan 26 09:32:09 crc kubenswrapper[4872]: I0126 09:32:09.193945 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eb2f617-9a5a-483c-aff5-3838bc2c68ec" path="/var/lib/kubelet/pods/9eb2f617-9a5a-483c-aff5-3838bc2c68ec/volumes" Jan 26 09:32:09 crc kubenswrapper[4872]: I0126 09:32:09.473278 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:32:09 crc kubenswrapper[4872]: I0126 09:32:09.473606 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:32:09 crc kubenswrapper[4872]: I0126 09:32:09.547003 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:32:10 crc kubenswrapper[4872]: I0126 09:32:10.405417 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:32:11 crc kubenswrapper[4872]: I0126 09:32:11.537539 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fxc65"] Jan 26 09:32:13 crc kubenswrapper[4872]: I0126 09:32:13.386904 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fxc65" podUID="336dfe89-6359-4d43-956f-8e9c119cf45e" containerName="registry-server" containerID="cri-o://6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409" gracePeriod=2 Jan 26 09:32:13 crc kubenswrapper[4872]: I0126 09:32:13.900817 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.104569 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/336dfe89-6359-4d43-956f-8e9c119cf45e-utilities\") pod \"336dfe89-6359-4d43-956f-8e9c119cf45e\" (UID: \"336dfe89-6359-4d43-956f-8e9c119cf45e\") " Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.110888 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqjjh\" (UniqueName: \"kubernetes.io/projected/336dfe89-6359-4d43-956f-8e9c119cf45e-kube-api-access-nqjjh\") pod \"336dfe89-6359-4d43-956f-8e9c119cf45e\" (UID: \"336dfe89-6359-4d43-956f-8e9c119cf45e\") " Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.105872 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/336dfe89-6359-4d43-956f-8e9c119cf45e-utilities" (OuterVolumeSpecName: "utilities") pod "336dfe89-6359-4d43-956f-8e9c119cf45e" (UID: "336dfe89-6359-4d43-956f-8e9c119cf45e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.111205 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/336dfe89-6359-4d43-956f-8e9c119cf45e-catalog-content\") pod \"336dfe89-6359-4d43-956f-8e9c119cf45e\" (UID: \"336dfe89-6359-4d43-956f-8e9c119cf45e\") " Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.112065 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/336dfe89-6359-4d43-956f-8e9c119cf45e-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.118471 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/336dfe89-6359-4d43-956f-8e9c119cf45e-kube-api-access-nqjjh" (OuterVolumeSpecName: "kube-api-access-nqjjh") pod "336dfe89-6359-4d43-956f-8e9c119cf45e" (UID: "336dfe89-6359-4d43-956f-8e9c119cf45e"). InnerVolumeSpecName "kube-api-access-nqjjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.164519 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/336dfe89-6359-4d43-956f-8e9c119cf45e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "336dfe89-6359-4d43-956f-8e9c119cf45e" (UID: "336dfe89-6359-4d43-956f-8e9c119cf45e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.213275 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqjjh\" (UniqueName: \"kubernetes.io/projected/336dfe89-6359-4d43-956f-8e9c119cf45e-kube-api-access-nqjjh\") on node \"crc\" DevicePath \"\"" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.213304 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/336dfe89-6359-4d43-956f-8e9c119cf45e-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.398251 4872 generic.go:334] "Generic (PLEG): container finished" podID="336dfe89-6359-4d43-956f-8e9c119cf45e" containerID="6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409" exitCode=0 Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.398360 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxc65" event={"ID":"336dfe89-6359-4d43-956f-8e9c119cf45e","Type":"ContainerDied","Data":"6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409"} Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.398469 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxc65" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.400581 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxc65" event={"ID":"336dfe89-6359-4d43-956f-8e9c119cf45e","Type":"ContainerDied","Data":"0adcd2e5f5d6d11cec42895422771396f0c880b3dbc835150cbf6904c62a2886"} Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.400925 4872 scope.go:117] "RemoveContainer" containerID="6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.428500 4872 scope.go:117] "RemoveContainer" containerID="68f1e2f5f9d9f9d7567d2882cbab053f9b1ec0a874b5df1bfd626c320963e6e5" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.470900 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fxc65"] Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.473725 4872 scope.go:117] "RemoveContainer" containerID="a95f31d902c0c641587bbf6f34859784acc03754639389862c0d97c82ab4124c" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.479764 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fxc65"] Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.528291 4872 scope.go:117] "RemoveContainer" containerID="6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409" Jan 26 09:32:14 crc kubenswrapper[4872]: E0126 09:32:14.528633 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409\": container with ID starting with 6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409 not found: ID does not exist" containerID="6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.528678 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409"} err="failed to get container status \"6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409\": rpc error: code = NotFound desc = could not find container \"6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409\": container with ID starting with 6f34237632ac6bfc6b6353dc236080adb71c1312f90b7df6b06e80bcd3b73409 not found: ID does not exist" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.528711 4872 scope.go:117] "RemoveContainer" containerID="68f1e2f5f9d9f9d7567d2882cbab053f9b1ec0a874b5df1bfd626c320963e6e5" Jan 26 09:32:14 crc kubenswrapper[4872]: E0126 09:32:14.529234 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68f1e2f5f9d9f9d7567d2882cbab053f9b1ec0a874b5df1bfd626c320963e6e5\": container with ID starting with 68f1e2f5f9d9f9d7567d2882cbab053f9b1ec0a874b5df1bfd626c320963e6e5 not found: ID does not exist" containerID="68f1e2f5f9d9f9d7567d2882cbab053f9b1ec0a874b5df1bfd626c320963e6e5" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.529263 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68f1e2f5f9d9f9d7567d2882cbab053f9b1ec0a874b5df1bfd626c320963e6e5"} err="failed to get container status \"68f1e2f5f9d9f9d7567d2882cbab053f9b1ec0a874b5df1bfd626c320963e6e5\": rpc error: code = NotFound desc = could not find container \"68f1e2f5f9d9f9d7567d2882cbab053f9b1ec0a874b5df1bfd626c320963e6e5\": container with ID starting with 68f1e2f5f9d9f9d7567d2882cbab053f9b1ec0a874b5df1bfd626c320963e6e5 not found: ID does not exist" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.529290 4872 scope.go:117] "RemoveContainer" containerID="a95f31d902c0c641587bbf6f34859784acc03754639389862c0d97c82ab4124c" Jan 26 09:32:14 crc kubenswrapper[4872]: E0126 09:32:14.529641 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a95f31d902c0c641587bbf6f34859784acc03754639389862c0d97c82ab4124c\": container with ID starting with a95f31d902c0c641587bbf6f34859784acc03754639389862c0d97c82ab4124c not found: ID does not exist" containerID="a95f31d902c0c641587bbf6f34859784acc03754639389862c0d97c82ab4124c" Jan 26 09:32:14 crc kubenswrapper[4872]: I0126 09:32:14.529691 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a95f31d902c0c641587bbf6f34859784acc03754639389862c0d97c82ab4124c"} err="failed to get container status \"a95f31d902c0c641587bbf6f34859784acc03754639389862c0d97c82ab4124c\": rpc error: code = NotFound desc = could not find container \"a95f31d902c0c641587bbf6f34859784acc03754639389862c0d97c82ab4124c\": container with ID starting with a95f31d902c0c641587bbf6f34859784acc03754639389862c0d97c82ab4124c not found: ID does not exist" Jan 26 09:32:15 crc kubenswrapper[4872]: I0126 09:32:15.198495 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="336dfe89-6359-4d43-956f-8e9c119cf45e" path="/var/lib/kubelet/pods/336dfe89-6359-4d43-956f-8e9c119cf45e/volumes" Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.059569 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-j82b4"] Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.071343 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-67f9-account-create-update-d2zpv"] Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.081769 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-j82b4"] Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.092480 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-x6t97"] Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.103960 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-f159-account-create-update-xxdzr"] Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.114633 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-76ba-account-create-update-4dklb"] Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.123310 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-h7qsd"] Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.133726 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-x6t97"] Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.143521 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-67f9-account-create-update-d2zpv"] Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.154142 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-f159-account-create-update-xxdzr"] Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.162449 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-h7qsd"] Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.219932 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0723de91-066f-474b-b7fb-f5f252a830bb" path="/var/lib/kubelet/pods/0723de91-066f-474b-b7fb-f5f252a830bb/volumes" Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.220732 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="392ab1c3-7cbf-44ac-8b8e-39b64e502d8a" path="/var/lib/kubelet/pods/392ab1c3-7cbf-44ac-8b8e-39b64e502d8a/volumes" Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.221384 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52cc682d-f3a9-47fc-9850-3019549c5f18" path="/var/lib/kubelet/pods/52cc682d-f3a9-47fc-9850-3019549c5f18/volumes" Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.222032 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="940d731a-47aa-40db-826b-7c1c7b23ee24" path="/var/lib/kubelet/pods/940d731a-47aa-40db-826b-7c1c7b23ee24/volumes" Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.223242 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfeb3d47-b023-4e41-9bdb-80241b5b0972" path="/var/lib/kubelet/pods/dfeb3d47-b023-4e41-9bdb-80241b5b0972/volumes" Jan 26 09:32:43 crc kubenswrapper[4872]: I0126 09:32:43.223856 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-76ba-account-create-update-4dklb"] Jan 26 09:32:45 crc kubenswrapper[4872]: I0126 09:32:45.196122 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1717069a-a49e-44ae-bb02-ebc6265e44b3" path="/var/lib/kubelet/pods/1717069a-a49e-44ae-bb02-ebc6265e44b3/volumes" Jan 26 09:33:11 crc kubenswrapper[4872]: I0126 09:33:11.050897 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-dnrdk"] Jan 26 09:33:11 crc kubenswrapper[4872]: I0126 09:33:11.062382 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-dnrdk"] Jan 26 09:33:11 crc kubenswrapper[4872]: I0126 09:33:11.277409 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01da7c35-0645-45eb-814b-7bfe35d14f46" path="/var/lib/kubelet/pods/01da7c35-0645-45eb-814b-7bfe35d14f46/volumes" Jan 26 09:33:15 crc kubenswrapper[4872]: I0126 09:33:15.210355 4872 scope.go:117] "RemoveContainer" containerID="b6c093bbea84de36a73de76ec30d497fa3c0bd401ed4f88003495408816cdeb4" Jan 26 09:33:15 crc kubenswrapper[4872]: I0126 09:33:15.260728 4872 scope.go:117] "RemoveContainer" containerID="9f1e344a48efae4a82a77a45f54bc9822ea9f569e4d9cd53d0c966e274ec5d9d" Jan 26 09:33:15 crc kubenswrapper[4872]: I0126 09:33:15.312519 4872 scope.go:117] "RemoveContainer" containerID="b6df7e8486927d214ef6662f821a83abaa172de17822d798e0e0c1bf0ef414f2" Jan 26 09:33:15 crc kubenswrapper[4872]: I0126 09:33:15.359314 4872 scope.go:117] "RemoveContainer" containerID="e449be6901dedad52713a04f03ce664d0afa53c4eb9ea0b060eb6574f4fd2223" Jan 26 09:33:15 crc kubenswrapper[4872]: I0126 09:33:15.421068 4872 scope.go:117] "RemoveContainer" containerID="fb54217da7329679731ceace5fdfe65aa5e1c5d9a7f57e23f7f53392e3580e8c" Jan 26 09:33:15 crc kubenswrapper[4872]: I0126 09:33:15.461540 4872 scope.go:117] "RemoveContainer" containerID="8efba7583a85ca87b0a3874892d9272bf4c2553acbc8b2acc196569532f77bc7" Jan 26 09:33:15 crc kubenswrapper[4872]: I0126 09:33:15.504360 4872 scope.go:117] "RemoveContainer" containerID="bf89441413031f9b78b00cf7b749940f7bccf809ad4cb47e9fb0a74c79b7b125" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.389363 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b5td7"] Jan 26 09:33:16 crc kubenswrapper[4872]: E0126 09:33:16.390386 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eb2f617-9a5a-483c-aff5-3838bc2c68ec" containerName="registry-server" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.390405 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eb2f617-9a5a-483c-aff5-3838bc2c68ec" containerName="registry-server" Jan 26 09:33:16 crc kubenswrapper[4872]: E0126 09:33:16.390424 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eb2f617-9a5a-483c-aff5-3838bc2c68ec" containerName="extract-utilities" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.390433 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eb2f617-9a5a-483c-aff5-3838bc2c68ec" containerName="extract-utilities" Jan 26 09:33:16 crc kubenswrapper[4872]: E0126 09:33:16.390451 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eb2f617-9a5a-483c-aff5-3838bc2c68ec" containerName="extract-content" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.390460 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eb2f617-9a5a-483c-aff5-3838bc2c68ec" containerName="extract-content" Jan 26 09:33:16 crc kubenswrapper[4872]: E0126 09:33:16.390475 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336dfe89-6359-4d43-956f-8e9c119cf45e" containerName="registry-server" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.390483 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="336dfe89-6359-4d43-956f-8e9c119cf45e" containerName="registry-server" Jan 26 09:33:16 crc kubenswrapper[4872]: E0126 09:33:16.390510 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336dfe89-6359-4d43-956f-8e9c119cf45e" containerName="extract-utilities" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.390518 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="336dfe89-6359-4d43-956f-8e9c119cf45e" containerName="extract-utilities" Jan 26 09:33:16 crc kubenswrapper[4872]: E0126 09:33:16.390532 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336dfe89-6359-4d43-956f-8e9c119cf45e" containerName="extract-content" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.390539 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="336dfe89-6359-4d43-956f-8e9c119cf45e" containerName="extract-content" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.390861 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="336dfe89-6359-4d43-956f-8e9c119cf45e" containerName="registry-server" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.390898 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eb2f617-9a5a-483c-aff5-3838bc2c68ec" containerName="registry-server" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.396400 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.411745 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b5td7"] Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.505316 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21a79119-dd83-4a50-b4ff-a1807bb2506d-utilities\") pod \"redhat-operators-b5td7\" (UID: \"21a79119-dd83-4a50-b4ff-a1807bb2506d\") " pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.505777 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98r8r\" (UniqueName: \"kubernetes.io/projected/21a79119-dd83-4a50-b4ff-a1807bb2506d-kube-api-access-98r8r\") pod \"redhat-operators-b5td7\" (UID: \"21a79119-dd83-4a50-b4ff-a1807bb2506d\") " pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.507135 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21a79119-dd83-4a50-b4ff-a1807bb2506d-catalog-content\") pod \"redhat-operators-b5td7\" (UID: \"21a79119-dd83-4a50-b4ff-a1807bb2506d\") " pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.609345 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21a79119-dd83-4a50-b4ff-a1807bb2506d-catalog-content\") pod \"redhat-operators-b5td7\" (UID: \"21a79119-dd83-4a50-b4ff-a1807bb2506d\") " pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.609447 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21a79119-dd83-4a50-b4ff-a1807bb2506d-utilities\") pod \"redhat-operators-b5td7\" (UID: \"21a79119-dd83-4a50-b4ff-a1807bb2506d\") " pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.609472 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98r8r\" (UniqueName: \"kubernetes.io/projected/21a79119-dd83-4a50-b4ff-a1807bb2506d-kube-api-access-98r8r\") pod \"redhat-operators-b5td7\" (UID: \"21a79119-dd83-4a50-b4ff-a1807bb2506d\") " pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.610520 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21a79119-dd83-4a50-b4ff-a1807bb2506d-utilities\") pod \"redhat-operators-b5td7\" (UID: \"21a79119-dd83-4a50-b4ff-a1807bb2506d\") " pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.610653 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21a79119-dd83-4a50-b4ff-a1807bb2506d-catalog-content\") pod \"redhat-operators-b5td7\" (UID: \"21a79119-dd83-4a50-b4ff-a1807bb2506d\") " pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.644503 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98r8r\" (UniqueName: \"kubernetes.io/projected/21a79119-dd83-4a50-b4ff-a1807bb2506d-kube-api-access-98r8r\") pod \"redhat-operators-b5td7\" (UID: \"21a79119-dd83-4a50-b4ff-a1807bb2506d\") " pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:16 crc kubenswrapper[4872]: I0126 09:33:16.745755 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:17 crc kubenswrapper[4872]: I0126 09:33:17.219439 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b5td7"] Jan 26 09:33:17 crc kubenswrapper[4872]: I0126 09:33:17.361413 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5td7" event={"ID":"21a79119-dd83-4a50-b4ff-a1807bb2506d","Type":"ContainerStarted","Data":"6878133905eb27a54deb43316d31fc0642a1e06b6bb67f17fd33eb15479d5dfe"} Jan 26 09:33:18 crc kubenswrapper[4872]: I0126 09:33:18.372647 4872 generic.go:334] "Generic (PLEG): container finished" podID="21a79119-dd83-4a50-b4ff-a1807bb2506d" containerID="f71b3c99a6e7ff4c985f75f577903b7ce0b7144623d8302127af3f7229bb5cdd" exitCode=0 Jan 26 09:33:18 crc kubenswrapper[4872]: I0126 09:33:18.372769 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5td7" event={"ID":"21a79119-dd83-4a50-b4ff-a1807bb2506d","Type":"ContainerDied","Data":"f71b3c99a6e7ff4c985f75f577903b7ce0b7144623d8302127af3f7229bb5cdd"} Jan 26 09:33:19 crc kubenswrapper[4872]: I0126 09:33:19.034539 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-v4kh2"] Jan 26 09:33:19 crc kubenswrapper[4872]: I0126 09:33:19.041244 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-v4kh2"] Jan 26 09:33:19 crc kubenswrapper[4872]: I0126 09:33:19.196292 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d" path="/var/lib/kubelet/pods/096e3ab6-99aa-47f4-8ecf-2f95d7d8fd5d/volumes" Jan 26 09:33:19 crc kubenswrapper[4872]: I0126 09:33:19.391989 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5td7" event={"ID":"21a79119-dd83-4a50-b4ff-a1807bb2506d","Type":"ContainerStarted","Data":"896e69ba9c3b97051a577b9163a5679912d6c31e00284fa98902af062e4dd4c5"} Jan 26 09:33:22 crc kubenswrapper[4872]: I0126 09:33:22.062783 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-24ac-account-create-update-vp5nc"] Jan 26 09:33:22 crc kubenswrapper[4872]: I0126 09:33:22.081644 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6421-account-create-update-75kn6"] Jan 26 09:33:22 crc kubenswrapper[4872]: I0126 09:33:22.093853 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-28cqb"] Jan 26 09:33:22 crc kubenswrapper[4872]: I0126 09:33:22.105051 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-24ac-account-create-update-vp5nc"] Jan 26 09:33:22 crc kubenswrapper[4872]: I0126 09:33:22.115494 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-28cqb"] Jan 26 09:33:22 crc kubenswrapper[4872]: I0126 09:33:22.124586 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-xf6vh"] Jan 26 09:33:22 crc kubenswrapper[4872]: I0126 09:33:22.138141 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-xf6vh"] Jan 26 09:33:22 crc kubenswrapper[4872]: I0126 09:33:22.153133 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6421-account-create-update-75kn6"] Jan 26 09:33:22 crc kubenswrapper[4872]: I0126 09:33:22.431104 4872 generic.go:334] "Generic (PLEG): container finished" podID="21a79119-dd83-4a50-b4ff-a1807bb2506d" containerID="896e69ba9c3b97051a577b9163a5679912d6c31e00284fa98902af062e4dd4c5" exitCode=0 Jan 26 09:33:22 crc kubenswrapper[4872]: I0126 09:33:22.431193 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5td7" event={"ID":"21a79119-dd83-4a50-b4ff-a1807bb2506d","Type":"ContainerDied","Data":"896e69ba9c3b97051a577b9163a5679912d6c31e00284fa98902af062e4dd4c5"} Jan 26 09:33:23 crc kubenswrapper[4872]: I0126 09:33:23.041535 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-0c9f-account-create-update-q688k"] Jan 26 09:33:23 crc kubenswrapper[4872]: I0126 09:33:23.057575 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-0c9f-account-create-update-q688k"] Jan 26 09:33:23 crc kubenswrapper[4872]: I0126 09:33:23.204939 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20add17d-dd15-47ff-8d0c-5c6ca089fc65" path="/var/lib/kubelet/pods/20add17d-dd15-47ff-8d0c-5c6ca089fc65/volumes" Jan 26 09:33:23 crc kubenswrapper[4872]: I0126 09:33:23.205696 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d5889bf-aa3d-4588-9d97-4c184bca015b" path="/var/lib/kubelet/pods/4d5889bf-aa3d-4588-9d97-4c184bca015b/volumes" Jan 26 09:33:23 crc kubenswrapper[4872]: I0126 09:33:23.206423 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8759676c-42c7-41ef-b71e-f9848dd84ceb" path="/var/lib/kubelet/pods/8759676c-42c7-41ef-b71e-f9848dd84ceb/volumes" Jan 26 09:33:23 crc kubenswrapper[4872]: I0126 09:33:23.207178 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc7c1498-7864-484d-b86a-2f9ba512bf13" path="/var/lib/kubelet/pods/cc7c1498-7864-484d-b86a-2f9ba512bf13/volumes" Jan 26 09:33:23 crc kubenswrapper[4872]: I0126 09:33:23.208506 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2a972ca-1c4f-4124-b481-0eeca0e1ac06" path="/var/lib/kubelet/pods/d2a972ca-1c4f-4124-b481-0eeca0e1ac06/volumes" Jan 26 09:33:23 crc kubenswrapper[4872]: I0126 09:33:23.446221 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5td7" event={"ID":"21a79119-dd83-4a50-b4ff-a1807bb2506d","Type":"ContainerStarted","Data":"155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c"} Jan 26 09:33:23 crc kubenswrapper[4872]: I0126 09:33:23.478411 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b5td7" podStartSLOduration=2.745911151 podStartE2EDuration="7.478379981s" podCreationTimestamp="2026-01-26 09:33:16 +0000 UTC" firstStartedPulling="2026-01-26 09:33:18.376602431 +0000 UTC m=+1531.685442252" lastFinishedPulling="2026-01-26 09:33:23.109071271 +0000 UTC m=+1536.417911082" observedRunningTime="2026-01-26 09:33:23.467851867 +0000 UTC m=+1536.776691668" watchObservedRunningTime="2026-01-26 09:33:23.478379981 +0000 UTC m=+1536.787219782" Jan 26 09:33:26 crc kubenswrapper[4872]: I0126 09:33:26.746033 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:26 crc kubenswrapper[4872]: I0126 09:33:26.747047 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:27 crc kubenswrapper[4872]: I0126 09:33:27.807984 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-b5td7" podUID="21a79119-dd83-4a50-b4ff-a1807bb2506d" containerName="registry-server" probeResult="failure" output=< Jan 26 09:33:27 crc kubenswrapper[4872]: timeout: failed to connect service ":50051" within 1s Jan 26 09:33:27 crc kubenswrapper[4872]: > Jan 26 09:33:28 crc kubenswrapper[4872]: I0126 09:33:28.039632 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-ccwl7"] Jan 26 09:33:28 crc kubenswrapper[4872]: I0126 09:33:28.053519 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-ccwl7"] Jan 26 09:33:29 crc kubenswrapper[4872]: I0126 09:33:29.195198 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3741809-0b4f-4816-9130-c9f8d73aaded" path="/var/lib/kubelet/pods/b3741809-0b4f-4816-9130-c9f8d73aaded/volumes" Jan 26 09:33:34 crc kubenswrapper[4872]: I0126 09:33:34.076233 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-tchz8"] Jan 26 09:33:34 crc kubenswrapper[4872]: I0126 09:33:34.092868 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-db-create-jdmsk"] Jan 26 09:33:34 crc kubenswrapper[4872]: I0126 09:33:34.103105 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-tchz8"] Jan 26 09:33:34 crc kubenswrapper[4872]: I0126 09:33:34.113128 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-c3e0-account-create-update-pfn6b"] Jan 26 09:33:34 crc kubenswrapper[4872]: I0126 09:33:34.121949 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-db-create-jdmsk"] Jan 26 09:33:34 crc kubenswrapper[4872]: I0126 09:33:34.130040 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-c3e0-account-create-update-pfn6b"] Jan 26 09:33:35 crc kubenswrapper[4872]: I0126 09:33:35.201171 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="342ea714-cff5-4b5b-bf68-f7baf40bd84e" path="/var/lib/kubelet/pods/342ea714-cff5-4b5b-bf68-f7baf40bd84e/volumes" Jan 26 09:33:35 crc kubenswrapper[4872]: I0126 09:33:35.203503 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00" path="/var/lib/kubelet/pods/abe6ebd7-4f3f-4626-ba1c-7bc41fdfed00/volumes" Jan 26 09:33:35 crc kubenswrapper[4872]: I0126 09:33:35.207104 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de9df0f9-9456-4bfd-98f7-065a933bb80e" path="/var/lib/kubelet/pods/de9df0f9-9456-4bfd-98f7-065a933bb80e/volumes" Jan 26 09:33:36 crc kubenswrapper[4872]: I0126 09:33:36.809511 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:36 crc kubenswrapper[4872]: I0126 09:33:36.884072 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:37 crc kubenswrapper[4872]: I0126 09:33:37.060849 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b5td7"] Jan 26 09:33:38 crc kubenswrapper[4872]: I0126 09:33:38.639604 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b5td7" podUID="21a79119-dd83-4a50-b4ff-a1807bb2506d" containerName="registry-server" containerID="cri-o://155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c" gracePeriod=2 Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.175339 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.248142 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21a79119-dd83-4a50-b4ff-a1807bb2506d-utilities\") pod \"21a79119-dd83-4a50-b4ff-a1807bb2506d\" (UID: \"21a79119-dd83-4a50-b4ff-a1807bb2506d\") " Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.248262 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98r8r\" (UniqueName: \"kubernetes.io/projected/21a79119-dd83-4a50-b4ff-a1807bb2506d-kube-api-access-98r8r\") pod \"21a79119-dd83-4a50-b4ff-a1807bb2506d\" (UID: \"21a79119-dd83-4a50-b4ff-a1807bb2506d\") " Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.248374 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21a79119-dd83-4a50-b4ff-a1807bb2506d-catalog-content\") pod \"21a79119-dd83-4a50-b4ff-a1807bb2506d\" (UID: \"21a79119-dd83-4a50-b4ff-a1807bb2506d\") " Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.249206 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21a79119-dd83-4a50-b4ff-a1807bb2506d-utilities" (OuterVolumeSpecName: "utilities") pod "21a79119-dd83-4a50-b4ff-a1807bb2506d" (UID: "21a79119-dd83-4a50-b4ff-a1807bb2506d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.250533 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/21a79119-dd83-4a50-b4ff-a1807bb2506d-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.266834 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21a79119-dd83-4a50-b4ff-a1807bb2506d-kube-api-access-98r8r" (OuterVolumeSpecName: "kube-api-access-98r8r") pod "21a79119-dd83-4a50-b4ff-a1807bb2506d" (UID: "21a79119-dd83-4a50-b4ff-a1807bb2506d"). InnerVolumeSpecName "kube-api-access-98r8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.353016 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98r8r\" (UniqueName: \"kubernetes.io/projected/21a79119-dd83-4a50-b4ff-a1807bb2506d-kube-api-access-98r8r\") on node \"crc\" DevicePath \"\"" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.420601 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21a79119-dd83-4a50-b4ff-a1807bb2506d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "21a79119-dd83-4a50-b4ff-a1807bb2506d" (UID: "21a79119-dd83-4a50-b4ff-a1807bb2506d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.455253 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/21a79119-dd83-4a50-b4ff-a1807bb2506d-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.658812 4872 generic.go:334] "Generic (PLEG): container finished" podID="21a79119-dd83-4a50-b4ff-a1807bb2506d" containerID="155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c" exitCode=0 Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.658865 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5td7" event={"ID":"21a79119-dd83-4a50-b4ff-a1807bb2506d","Type":"ContainerDied","Data":"155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c"} Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.658900 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b5td7" event={"ID":"21a79119-dd83-4a50-b4ff-a1807bb2506d","Type":"ContainerDied","Data":"6878133905eb27a54deb43316d31fc0642a1e06b6bb67f17fd33eb15479d5dfe"} Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.658923 4872 scope.go:117] "RemoveContainer" containerID="155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.659104 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b5td7" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.711829 4872 scope.go:117] "RemoveContainer" containerID="896e69ba9c3b97051a577b9163a5679912d6c31e00284fa98902af062e4dd4c5" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.726633 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b5td7"] Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.741230 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b5td7"] Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.759772 4872 scope.go:117] "RemoveContainer" containerID="f71b3c99a6e7ff4c985f75f577903b7ce0b7144623d8302127af3f7229bb5cdd" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.812948 4872 scope.go:117] "RemoveContainer" containerID="155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c" Jan 26 09:33:39 crc kubenswrapper[4872]: E0126 09:33:39.813468 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c\": container with ID starting with 155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c not found: ID does not exist" containerID="155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.813508 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c"} err="failed to get container status \"155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c\": rpc error: code = NotFound desc = could not find container \"155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c\": container with ID starting with 155175c403f83bf6c8790602e3f740320fea683d386ce766a57f9a293296391c not found: ID does not exist" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.813550 4872 scope.go:117] "RemoveContainer" containerID="896e69ba9c3b97051a577b9163a5679912d6c31e00284fa98902af062e4dd4c5" Jan 26 09:33:39 crc kubenswrapper[4872]: E0126 09:33:39.813833 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"896e69ba9c3b97051a577b9163a5679912d6c31e00284fa98902af062e4dd4c5\": container with ID starting with 896e69ba9c3b97051a577b9163a5679912d6c31e00284fa98902af062e4dd4c5 not found: ID does not exist" containerID="896e69ba9c3b97051a577b9163a5679912d6c31e00284fa98902af062e4dd4c5" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.813883 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"896e69ba9c3b97051a577b9163a5679912d6c31e00284fa98902af062e4dd4c5"} err="failed to get container status \"896e69ba9c3b97051a577b9163a5679912d6c31e00284fa98902af062e4dd4c5\": rpc error: code = NotFound desc = could not find container \"896e69ba9c3b97051a577b9163a5679912d6c31e00284fa98902af062e4dd4c5\": container with ID starting with 896e69ba9c3b97051a577b9163a5679912d6c31e00284fa98902af062e4dd4c5 not found: ID does not exist" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.813897 4872 scope.go:117] "RemoveContainer" containerID="f71b3c99a6e7ff4c985f75f577903b7ce0b7144623d8302127af3f7229bb5cdd" Jan 26 09:33:39 crc kubenswrapper[4872]: E0126 09:33:39.814150 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f71b3c99a6e7ff4c985f75f577903b7ce0b7144623d8302127af3f7229bb5cdd\": container with ID starting with f71b3c99a6e7ff4c985f75f577903b7ce0b7144623d8302127af3f7229bb5cdd not found: ID does not exist" containerID="f71b3c99a6e7ff4c985f75f577903b7ce0b7144623d8302127af3f7229bb5cdd" Jan 26 09:33:39 crc kubenswrapper[4872]: I0126 09:33:39.814173 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f71b3c99a6e7ff4c985f75f577903b7ce0b7144623d8302127af3f7229bb5cdd"} err="failed to get container status \"f71b3c99a6e7ff4c985f75f577903b7ce0b7144623d8302127af3f7229bb5cdd\": rpc error: code = NotFound desc = could not find container \"f71b3c99a6e7ff4c985f75f577903b7ce0b7144623d8302127af3f7229bb5cdd\": container with ID starting with f71b3c99a6e7ff4c985f75f577903b7ce0b7144623d8302127af3f7229bb5cdd not found: ID does not exist" Jan 26 09:33:40 crc kubenswrapper[4872]: I0126 09:33:40.227618 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:33:40 crc kubenswrapper[4872]: I0126 09:33:40.228124 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:33:41 crc kubenswrapper[4872]: I0126 09:33:41.205192 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21a79119-dd83-4a50-b4ff-a1807bb2506d" path="/var/lib/kubelet/pods/21a79119-dd83-4a50-b4ff-a1807bb2506d/volumes" Jan 26 09:34:02 crc kubenswrapper[4872]: I0126 09:34:02.086438 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-zdqwd"] Jan 26 09:34:02 crc kubenswrapper[4872]: I0126 09:34:02.107497 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-p27pg"] Jan 26 09:34:02 crc kubenswrapper[4872]: I0126 09:34:02.124613 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-zd82w"] Jan 26 09:34:02 crc kubenswrapper[4872]: I0126 09:34:02.136726 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-p27pg"] Jan 26 09:34:02 crc kubenswrapper[4872]: I0126 09:34:02.148398 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-zdqwd"] Jan 26 09:34:02 crc kubenswrapper[4872]: I0126 09:34:02.160295 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-zd82w"] Jan 26 09:34:03 crc kubenswrapper[4872]: I0126 09:34:03.205134 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04076889-57e8-41ad-b47e-6d8e5e23223f" path="/var/lib/kubelet/pods/04076889-57e8-41ad-b47e-6d8e5e23223f/volumes" Jan 26 09:34:03 crc kubenswrapper[4872]: I0126 09:34:03.205941 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a9d0b8b-dad6-4dbd-a261-77264014fc9b" path="/var/lib/kubelet/pods/0a9d0b8b-dad6-4dbd-a261-77264014fc9b/volumes" Jan 26 09:34:03 crc kubenswrapper[4872]: I0126 09:34:03.206609 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b" path="/var/lib/kubelet/pods/1f6ad3ad-8b9d-4dc0-bf7a-7b7eff1e666b/volumes" Jan 26 09:34:10 crc kubenswrapper[4872]: I0126 09:34:10.227468 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:34:10 crc kubenswrapper[4872]: I0126 09:34:10.228121 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:34:15 crc kubenswrapper[4872]: I0126 09:34:15.707295 4872 scope.go:117] "RemoveContainer" containerID="3eab026336d335767282c539c580990b865721784af21ad1996aa17c6a95c0d4" Jan 26 09:34:15 crc kubenswrapper[4872]: I0126 09:34:15.754691 4872 scope.go:117] "RemoveContainer" containerID="99a0d1c30dd823ab40c90826b5604682366f97f920ee24739cd9b0346a6f22bd" Jan 26 09:34:15 crc kubenswrapper[4872]: I0126 09:34:15.844598 4872 scope.go:117] "RemoveContainer" containerID="f827292a70bfdbade694480b6f9973315117fbf091e98aa6a6ab496f2aaa0b15" Jan 26 09:34:15 crc kubenswrapper[4872]: I0126 09:34:15.877436 4872 scope.go:117] "RemoveContainer" containerID="8c0611984b881bce81eac16dd33b546dba9189a4544cd647d0da16b009a27346" Jan 26 09:34:15 crc kubenswrapper[4872]: I0126 09:34:15.927356 4872 scope.go:117] "RemoveContainer" containerID="285157d6fe4d1fdddbc90cb5c3d52d6ea2d827a0cecf5ac04d4964503cedc339" Jan 26 09:34:15 crc kubenswrapper[4872]: I0126 09:34:15.980493 4872 scope.go:117] "RemoveContainer" containerID="b27def704e2ed5900d0a9e45a5bff4f0c1e906e0031f7970dd6c733e48aaa678" Jan 26 09:34:16 crc kubenswrapper[4872]: I0126 09:34:16.017640 4872 scope.go:117] "RemoveContainer" containerID="6edb450c51c6475eee8f7ffd305e21c37ca6b8ec3c74ce39fd72f7e282bea9b2" Jan 26 09:34:16 crc kubenswrapper[4872]: I0126 09:34:16.042688 4872 scope.go:117] "RemoveContainer" containerID="7f97ad1a95a364f7c129725296d2cdac6e3572e58dc16246927a099f72bf3f66" Jan 26 09:34:16 crc kubenswrapper[4872]: I0126 09:34:16.071255 4872 scope.go:117] "RemoveContainer" containerID="958517ad3eeeb9238b0d2f145a6ce26ae1a0127d403f1808b713aa7b5df84638" Jan 26 09:34:16 crc kubenswrapper[4872]: I0126 09:34:16.122319 4872 scope.go:117] "RemoveContainer" containerID="32b2ef75c3bf80927edfec877a66d09ed6e86666394fdb8b5858bf3dae6e5205" Jan 26 09:34:16 crc kubenswrapper[4872]: I0126 09:34:16.154249 4872 scope.go:117] "RemoveContainer" containerID="14c4e3f971586e364f1d40ef94a38c377b3983ed0b4ecf9f6125b69448b454d4" Jan 26 09:34:16 crc kubenswrapper[4872]: I0126 09:34:16.203828 4872 scope.go:117] "RemoveContainer" containerID="eddd753e23eec8e2b29820b3ed6a10bea1f8da1dc5b84bf2954290cbb853e838" Jan 26 09:34:16 crc kubenswrapper[4872]: I0126 09:34:16.248190 4872 scope.go:117] "RemoveContainer" containerID="ad4e458c514b7df57ac246f939d7104cfd24233db841179b21dd81517401bbc0" Jan 26 09:34:19 crc kubenswrapper[4872]: I0126 09:34:19.049274 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-dk64w"] Jan 26 09:34:19 crc kubenswrapper[4872]: I0126 09:34:19.065209 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-dk64w"] Jan 26 09:34:19 crc kubenswrapper[4872]: I0126 09:34:19.207284 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddd0e60e-a07c-41b0-8c44-00844dfef24a" path="/var/lib/kubelet/pods/ddd0e60e-a07c-41b0-8c44-00844dfef24a/volumes" Jan 26 09:34:20 crc kubenswrapper[4872]: I0126 09:34:20.041338 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-4cbzl"] Jan 26 09:34:20 crc kubenswrapper[4872]: I0126 09:34:20.054135 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-4cbzl"] Jan 26 09:34:21 crc kubenswrapper[4872]: I0126 09:34:21.206373 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f653549-36ce-4db8-9c67-ec95b628e867" path="/var/lib/kubelet/pods/3f653549-36ce-4db8-9c67-ec95b628e867/volumes" Jan 26 09:34:28 crc kubenswrapper[4872]: I0126 09:34:28.042337 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-b7e6-account-create-update-mdh5l"] Jan 26 09:34:28 crc kubenswrapper[4872]: I0126 09:34:28.060911 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-b7e6-account-create-update-mdh5l"] Jan 26 09:34:28 crc kubenswrapper[4872]: I0126 09:34:28.073075 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-db-create-mhtml"] Jan 26 09:34:28 crc kubenswrapper[4872]: I0126 09:34:28.110776 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-db-create-mhtml"] Jan 26 09:34:29 crc kubenswrapper[4872]: I0126 09:34:29.199521 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a7aa3be-d2ee-415a-87ac-bb65176dee59" path="/var/lib/kubelet/pods/8a7aa3be-d2ee-415a-87ac-bb65176dee59/volumes" Jan 26 09:34:29 crc kubenswrapper[4872]: I0126 09:34:29.200742 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2" path="/var/lib/kubelet/pods/cae3b7a3-e06c-4f8c-b54a-b3585e4e13f2/volumes" Jan 26 09:34:40 crc kubenswrapper[4872]: I0126 09:34:40.227560 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:34:40 crc kubenswrapper[4872]: I0126 09:34:40.228265 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:34:40 crc kubenswrapper[4872]: I0126 09:34:40.228330 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:34:40 crc kubenswrapper[4872]: I0126 09:34:40.229149 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a"} pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 09:34:40 crc kubenswrapper[4872]: I0126 09:34:40.229237 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" containerID="cri-o://fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" gracePeriod=600 Jan 26 09:34:40 crc kubenswrapper[4872]: E0126 09:34:40.369602 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:34:41 crc kubenswrapper[4872]: I0126 09:34:41.374991 4872 generic.go:334] "Generic (PLEG): container finished" podID="fca44d96-a000-4bf2-8283-a937b0192880" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" exitCode=0 Jan 26 09:34:41 crc kubenswrapper[4872]: I0126 09:34:41.375084 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerDied","Data":"fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a"} Jan 26 09:34:41 crc kubenswrapper[4872]: I0126 09:34:41.375620 4872 scope.go:117] "RemoveContainer" containerID="16869c8c2bd19091ae76a4591e6a82aaa7c0510c4b2bb4b199056a62beb5e880" Jan 26 09:34:41 crc kubenswrapper[4872]: I0126 09:34:41.376529 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:34:41 crc kubenswrapper[4872]: E0126 09:34:41.376843 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:34:52 crc kubenswrapper[4872]: I0126 09:34:52.183971 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:34:52 crc kubenswrapper[4872]: E0126 09:34:52.184904 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:35:05 crc kubenswrapper[4872]: I0126 09:35:05.184824 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:35:05 crc kubenswrapper[4872]: E0126 09:35:05.185731 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:35:11 crc kubenswrapper[4872]: I0126 09:35:11.053579 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-bkbz5"] Jan 26 09:35:11 crc kubenswrapper[4872]: I0126 09:35:11.063000 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-bkbz5"] Jan 26 09:35:11 crc kubenswrapper[4872]: I0126 09:35:11.078499 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-5935-account-create-update-ftjg4"] Jan 26 09:35:11 crc kubenswrapper[4872]: I0126 09:35:11.092364 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-5935-account-create-update-ftjg4"] Jan 26 09:35:11 crc kubenswrapper[4872]: I0126 09:35:11.197076 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24282d99-b526-4bd3-96b3-653642c42b9c" path="/var/lib/kubelet/pods/24282d99-b526-4bd3-96b3-653642c42b9c/volumes" Jan 26 09:35:11 crc kubenswrapper[4872]: I0126 09:35:11.197877 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d61a34c-59b3-490d-94a2-b9186426f179" path="/var/lib/kubelet/pods/8d61a34c-59b3-490d-94a2-b9186426f179/volumes" Jan 26 09:35:12 crc kubenswrapper[4872]: I0126 09:35:12.047944 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-f91e-account-create-update-twns5"] Jan 26 09:35:12 crc kubenswrapper[4872]: I0126 09:35:12.065393 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-6dpv7"] Jan 26 09:35:12 crc kubenswrapper[4872]: I0126 09:35:12.075540 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-cl2gt"] Jan 26 09:35:12 crc kubenswrapper[4872]: I0126 09:35:12.084724 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-4bea-account-create-update-vh8qs"] Jan 26 09:35:12 crc kubenswrapper[4872]: I0126 09:35:12.093619 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-f91e-account-create-update-twns5"] Jan 26 09:35:12 crc kubenswrapper[4872]: I0126 09:35:12.102739 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-6dpv7"] Jan 26 09:35:12 crc kubenswrapper[4872]: I0126 09:35:12.108640 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-4bea-account-create-update-vh8qs"] Jan 26 09:35:12 crc kubenswrapper[4872]: I0126 09:35:12.114582 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-cl2gt"] Jan 26 09:35:13 crc kubenswrapper[4872]: I0126 09:35:13.201016 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06fe0f9a-dbc9-4078-92ce-07fe06eb4110" path="/var/lib/kubelet/pods/06fe0f9a-dbc9-4078-92ce-07fe06eb4110/volumes" Jan 26 09:35:13 crc kubenswrapper[4872]: I0126 09:35:13.202515 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff" path="/var/lib/kubelet/pods/168518f2-b9e1-4d3f-a5fc-4b03f7ba50ff/volumes" Jan 26 09:35:13 crc kubenswrapper[4872]: I0126 09:35:13.203369 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f99a167-975f-4373-8c9d-23a2ab6510fa" path="/var/lib/kubelet/pods/4f99a167-975f-4373-8c9d-23a2ab6510fa/volumes" Jan 26 09:35:13 crc kubenswrapper[4872]: I0126 09:35:13.204015 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="574bcef8-68d6-40d8-adbb-5a321a3b59ce" path="/var/lib/kubelet/pods/574bcef8-68d6-40d8-adbb-5a321a3b59ce/volumes" Jan 26 09:35:16 crc kubenswrapper[4872]: I0126 09:35:16.628560 4872 scope.go:117] "RemoveContainer" containerID="7025f66c80d638b994c0a2de1d945d0fc7e34a9119ce3a2e383de1d4b7dd9285" Jan 26 09:35:16 crc kubenswrapper[4872]: I0126 09:35:16.680669 4872 scope.go:117] "RemoveContainer" containerID="7c32901cada88266eae0b24839df8e2f9fbbb73d60edebc33f736b085cb61ecf" Jan 26 09:35:16 crc kubenswrapper[4872]: I0126 09:35:16.729267 4872 scope.go:117] "RemoveContainer" containerID="427462d4b3684a01399799fb4af8b2bcd3411a3b2bd21f4c7b8e83d49a115241" Jan 26 09:35:16 crc kubenswrapper[4872]: I0126 09:35:16.791435 4872 scope.go:117] "RemoveContainer" containerID="7b91ffb556bd6ff0114ba45425a3e34c8ebd6365489b8787ecaca6cbf570da49" Jan 26 09:35:16 crc kubenswrapper[4872]: I0126 09:35:16.823541 4872 scope.go:117] "RemoveContainer" containerID="ee02f06a56b72178f0d9662be06e1101af2d56c208359c4dd1c813b29bdd473e" Jan 26 09:35:16 crc kubenswrapper[4872]: I0126 09:35:16.868136 4872 scope.go:117] "RemoveContainer" containerID="c55680258422d2aa3b674b16796ad6a3f6e39924b283c61962d5e34b4eeb6390" Jan 26 09:35:16 crc kubenswrapper[4872]: I0126 09:35:16.907334 4872 scope.go:117] "RemoveContainer" containerID="34f22c5846eb4b42bbf6fc56fe6e362757e1015f73a401c5bd8a46fcd3cab25b" Jan 26 09:35:16 crc kubenswrapper[4872]: I0126 09:35:16.928092 4872 scope.go:117] "RemoveContainer" containerID="b057a2f81ec82a726a1722537c79e5488dd003bd9df3b4017ede5e83a731656e" Jan 26 09:35:16 crc kubenswrapper[4872]: I0126 09:35:16.951862 4872 scope.go:117] "RemoveContainer" containerID="f829bcb22386227b578ea3bcc7d4d268f4fde0ea4a55e5965948a8f04318bafb" Jan 26 09:35:16 crc kubenswrapper[4872]: I0126 09:35:16.970497 4872 scope.go:117] "RemoveContainer" containerID="268730e4559ac67065bafe322e8335b0f075f20476100d6969b3ac265d2fedb9" Jan 26 09:35:20 crc kubenswrapper[4872]: I0126 09:35:20.184405 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:35:20 crc kubenswrapper[4872]: E0126 09:35:20.185015 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:35:35 crc kubenswrapper[4872]: I0126 09:35:35.185260 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:35:35 crc kubenswrapper[4872]: E0126 09:35:35.185877 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:35:42 crc kubenswrapper[4872]: I0126 09:35:42.050658 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ws4hm"] Jan 26 09:35:42 crc kubenswrapper[4872]: I0126 09:35:42.059205 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-ws4hm"] Jan 26 09:35:43 crc kubenswrapper[4872]: I0126 09:35:43.203936 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f45318ef-cb2f-4816-b4c4-360b1f446f5c" path="/var/lib/kubelet/pods/f45318ef-cb2f-4816-b4c4-360b1f446f5c/volumes" Jan 26 09:35:50 crc kubenswrapper[4872]: I0126 09:35:50.184142 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:35:50 crc kubenswrapper[4872]: E0126 09:35:50.185179 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:36:01 crc kubenswrapper[4872]: I0126 09:36:01.184098 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:36:01 crc kubenswrapper[4872]: E0126 09:36:01.185441 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:36:08 crc kubenswrapper[4872]: I0126 09:36:08.058283 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-srx6m"] Jan 26 09:36:08 crc kubenswrapper[4872]: I0126 09:36:08.075018 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-srx6m"] Jan 26 09:36:09 crc kubenswrapper[4872]: I0126 09:36:09.037005 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-449lj"] Jan 26 09:36:09 crc kubenswrapper[4872]: I0126 09:36:09.059599 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-449lj"] Jan 26 09:36:09 crc kubenswrapper[4872]: I0126 09:36:09.195880 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="384d42e1-70a6-44d9-9c4d-9862be13e0ce" path="/var/lib/kubelet/pods/384d42e1-70a6-44d9-9c4d-9862be13e0ce/volumes" Jan 26 09:36:09 crc kubenswrapper[4872]: I0126 09:36:09.196469 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e23cba8d-bf3b-4683-8f6d-ee2a29db6685" path="/var/lib/kubelet/pods/e23cba8d-bf3b-4683-8f6d-ee2a29db6685/volumes" Jan 26 09:36:12 crc kubenswrapper[4872]: I0126 09:36:12.184068 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:36:12 crc kubenswrapper[4872]: E0126 09:36:12.185206 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:36:17 crc kubenswrapper[4872]: I0126 09:36:17.238599 4872 scope.go:117] "RemoveContainer" containerID="3c19f590cae21f5eb10846e7021b9c69d79031b1800ae0a641cd0b8e69960457" Jan 26 09:36:17 crc kubenswrapper[4872]: I0126 09:36:17.282477 4872 scope.go:117] "RemoveContainer" containerID="1bb465fcc2f76c703eac83dc4771aa923acb2c20161bb28419cb57ad318181ab" Jan 26 09:36:17 crc kubenswrapper[4872]: I0126 09:36:17.345643 4872 scope.go:117] "RemoveContainer" containerID="20e0d61b49f43753f36008fd14e0e79c031c892fc01a6e8baf748d22ba044c71" Jan 26 09:36:27 crc kubenswrapper[4872]: I0126 09:36:27.191579 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:36:27 crc kubenswrapper[4872]: E0126 09:36:27.192639 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:36:42 crc kubenswrapper[4872]: I0126 09:36:42.185501 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:36:42 crc kubenswrapper[4872]: E0126 09:36:42.187141 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:36:52 crc kubenswrapper[4872]: I0126 09:36:52.062398 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-rmm5v"] Jan 26 09:36:52 crc kubenswrapper[4872]: I0126 09:36:52.078082 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-rmm5v"] Jan 26 09:36:53 crc kubenswrapper[4872]: I0126 09:36:53.204869 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3eec5e98-7d8f-4ee7-9bb8-e6824f931b01" path="/var/lib/kubelet/pods/3eec5e98-7d8f-4ee7-9bb8-e6824f931b01/volumes" Jan 26 09:36:56 crc kubenswrapper[4872]: I0126 09:36:56.185610 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:36:56 crc kubenswrapper[4872]: E0126 09:36:56.186711 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.655416 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6kdtd/must-gather-wfp8g"] Jan 26 09:37:00 crc kubenswrapper[4872]: E0126 09:37:00.659125 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a79119-dd83-4a50-b4ff-a1807bb2506d" containerName="extract-content" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.659249 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a79119-dd83-4a50-b4ff-a1807bb2506d" containerName="extract-content" Jan 26 09:37:00 crc kubenswrapper[4872]: E0126 09:37:00.659346 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a79119-dd83-4a50-b4ff-a1807bb2506d" containerName="extract-utilities" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.659436 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a79119-dd83-4a50-b4ff-a1807bb2506d" containerName="extract-utilities" Jan 26 09:37:00 crc kubenswrapper[4872]: E0126 09:37:00.659548 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a79119-dd83-4a50-b4ff-a1807bb2506d" containerName="registry-server" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.659636 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a79119-dd83-4a50-b4ff-a1807bb2506d" containerName="registry-server" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.659969 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="21a79119-dd83-4a50-b4ff-a1807bb2506d" containerName="registry-server" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.661378 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/must-gather-wfp8g" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.664014 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6kdtd"/"openshift-service-ca.crt" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.677783 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6kdtd/must-gather-wfp8g"] Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.688965 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-6kdtd"/"kube-root-ca.crt" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.751869 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/273d271b-605b-4dd7-b4d1-4046f63cef78-must-gather-output\") pod \"must-gather-wfp8g\" (UID: \"273d271b-605b-4dd7-b4d1-4046f63cef78\") " pod="openshift-must-gather-6kdtd/must-gather-wfp8g" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.751957 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z47tk\" (UniqueName: \"kubernetes.io/projected/273d271b-605b-4dd7-b4d1-4046f63cef78-kube-api-access-z47tk\") pod \"must-gather-wfp8g\" (UID: \"273d271b-605b-4dd7-b4d1-4046f63cef78\") " pod="openshift-must-gather-6kdtd/must-gather-wfp8g" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.853886 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/273d271b-605b-4dd7-b4d1-4046f63cef78-must-gather-output\") pod \"must-gather-wfp8g\" (UID: \"273d271b-605b-4dd7-b4d1-4046f63cef78\") " pod="openshift-must-gather-6kdtd/must-gather-wfp8g" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.854282 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z47tk\" (UniqueName: \"kubernetes.io/projected/273d271b-605b-4dd7-b4d1-4046f63cef78-kube-api-access-z47tk\") pod \"must-gather-wfp8g\" (UID: \"273d271b-605b-4dd7-b4d1-4046f63cef78\") " pod="openshift-must-gather-6kdtd/must-gather-wfp8g" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.854378 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/273d271b-605b-4dd7-b4d1-4046f63cef78-must-gather-output\") pod \"must-gather-wfp8g\" (UID: \"273d271b-605b-4dd7-b4d1-4046f63cef78\") " pod="openshift-must-gather-6kdtd/must-gather-wfp8g" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.875041 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z47tk\" (UniqueName: \"kubernetes.io/projected/273d271b-605b-4dd7-b4d1-4046f63cef78-kube-api-access-z47tk\") pod \"must-gather-wfp8g\" (UID: \"273d271b-605b-4dd7-b4d1-4046f63cef78\") " pod="openshift-must-gather-6kdtd/must-gather-wfp8g" Jan 26 09:37:00 crc kubenswrapper[4872]: I0126 09:37:00.994705 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/must-gather-wfp8g" Jan 26 09:37:01 crc kubenswrapper[4872]: I0126 09:37:01.527534 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-6kdtd/must-gather-wfp8g"] Jan 26 09:37:01 crc kubenswrapper[4872]: W0126 09:37:01.544165 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod273d271b_605b_4dd7_b4d1_4046f63cef78.slice/crio-5447112969a6f289f742275cbbd3e356e175b80b32376cb930a38ad77330610a WatchSource:0}: Error finding container 5447112969a6f289f742275cbbd3e356e175b80b32376cb930a38ad77330610a: Status 404 returned error can't find the container with id 5447112969a6f289f742275cbbd3e356e175b80b32376cb930a38ad77330610a Jan 26 09:37:01 crc kubenswrapper[4872]: I0126 09:37:01.548137 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 09:37:01 crc kubenswrapper[4872]: I0126 09:37:01.800144 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6kdtd/must-gather-wfp8g" event={"ID":"273d271b-605b-4dd7-b4d1-4046f63cef78","Type":"ContainerStarted","Data":"5447112969a6f289f742275cbbd3e356e175b80b32376cb930a38ad77330610a"} Jan 26 09:37:08 crc kubenswrapper[4872]: I0126 09:37:08.863525 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6kdtd/must-gather-wfp8g" event={"ID":"273d271b-605b-4dd7-b4d1-4046f63cef78","Type":"ContainerStarted","Data":"fb717a9caf98e4c05515f00cdc6e43d955a9a62f1062688a548d89f2dcaeb5d9"} Jan 26 09:37:08 crc kubenswrapper[4872]: I0126 09:37:08.864214 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6kdtd/must-gather-wfp8g" event={"ID":"273d271b-605b-4dd7-b4d1-4046f63cef78","Type":"ContainerStarted","Data":"c08c21452faf92710d7919aae65fdb95a33085480e9d7020c8deac1be4b76925"} Jan 26 09:37:08 crc kubenswrapper[4872]: I0126 09:37:08.892483 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6kdtd/must-gather-wfp8g" podStartSLOduration=2.430301888 podStartE2EDuration="8.892458708s" podCreationTimestamp="2026-01-26 09:37:00 +0000 UTC" firstStartedPulling="2026-01-26 09:37:01.548060763 +0000 UTC m=+1754.856900584" lastFinishedPulling="2026-01-26 09:37:08.010217603 +0000 UTC m=+1761.319057404" observedRunningTime="2026-01-26 09:37:08.887331386 +0000 UTC m=+1762.196171197" watchObservedRunningTime="2026-01-26 09:37:08.892458708 +0000 UTC m=+1762.201298509" Jan 26 09:37:11 crc kubenswrapper[4872]: E0126 09:37:11.094245 4872 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.70:41806->38.102.83.70:38695: write tcp 38.102.83.70:41806->38.102.83.70:38695: write: broken pipe Jan 26 09:37:11 crc kubenswrapper[4872]: I0126 09:37:11.183584 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:37:11 crc kubenswrapper[4872]: E0126 09:37:11.183844 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:37:11 crc kubenswrapper[4872]: I0126 09:37:11.884577 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6kdtd/crc-debug-cd879"] Jan 26 09:37:11 crc kubenswrapper[4872]: I0126 09:37:11.886079 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/crc-debug-cd879" Jan 26 09:37:11 crc kubenswrapper[4872]: I0126 09:37:11.888637 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-6kdtd"/"default-dockercfg-r2hgb" Jan 26 09:37:11 crc kubenswrapper[4872]: I0126 09:37:11.997153 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5dt7\" (UniqueName: \"kubernetes.io/projected/2b6bd18e-9143-4c5d-b08b-9b497271d5a8-kube-api-access-m5dt7\") pod \"crc-debug-cd879\" (UID: \"2b6bd18e-9143-4c5d-b08b-9b497271d5a8\") " pod="openshift-must-gather-6kdtd/crc-debug-cd879" Jan 26 09:37:11 crc kubenswrapper[4872]: I0126 09:37:11.997435 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b6bd18e-9143-4c5d-b08b-9b497271d5a8-host\") pod \"crc-debug-cd879\" (UID: \"2b6bd18e-9143-4c5d-b08b-9b497271d5a8\") " pod="openshift-must-gather-6kdtd/crc-debug-cd879" Jan 26 09:37:12 crc kubenswrapper[4872]: I0126 09:37:12.099619 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b6bd18e-9143-4c5d-b08b-9b497271d5a8-host\") pod \"crc-debug-cd879\" (UID: \"2b6bd18e-9143-4c5d-b08b-9b497271d5a8\") " pod="openshift-must-gather-6kdtd/crc-debug-cd879" Jan 26 09:37:12 crc kubenswrapper[4872]: I0126 09:37:12.099739 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5dt7\" (UniqueName: \"kubernetes.io/projected/2b6bd18e-9143-4c5d-b08b-9b497271d5a8-kube-api-access-m5dt7\") pod \"crc-debug-cd879\" (UID: \"2b6bd18e-9143-4c5d-b08b-9b497271d5a8\") " pod="openshift-must-gather-6kdtd/crc-debug-cd879" Jan 26 09:37:12 crc kubenswrapper[4872]: I0126 09:37:12.099845 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b6bd18e-9143-4c5d-b08b-9b497271d5a8-host\") pod \"crc-debug-cd879\" (UID: \"2b6bd18e-9143-4c5d-b08b-9b497271d5a8\") " pod="openshift-must-gather-6kdtd/crc-debug-cd879" Jan 26 09:37:12 crc kubenswrapper[4872]: I0126 09:37:12.123887 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5dt7\" (UniqueName: \"kubernetes.io/projected/2b6bd18e-9143-4c5d-b08b-9b497271d5a8-kube-api-access-m5dt7\") pod \"crc-debug-cd879\" (UID: \"2b6bd18e-9143-4c5d-b08b-9b497271d5a8\") " pod="openshift-must-gather-6kdtd/crc-debug-cd879" Jan 26 09:37:12 crc kubenswrapper[4872]: I0126 09:37:12.223460 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/crc-debug-cd879" Jan 26 09:37:12 crc kubenswrapper[4872]: W0126 09:37:12.271234 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b6bd18e_9143_4c5d_b08b_9b497271d5a8.slice/crio-f726c2e013e9aab978c85b9f94f1d2927aca0c3497cd36396fdde0a4cff41848 WatchSource:0}: Error finding container f726c2e013e9aab978c85b9f94f1d2927aca0c3497cd36396fdde0a4cff41848: Status 404 returned error can't find the container with id f726c2e013e9aab978c85b9f94f1d2927aca0c3497cd36396fdde0a4cff41848 Jan 26 09:37:12 crc kubenswrapper[4872]: I0126 09:37:12.933607 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6kdtd/crc-debug-cd879" event={"ID":"2b6bd18e-9143-4c5d-b08b-9b497271d5a8","Type":"ContainerStarted","Data":"f726c2e013e9aab978c85b9f94f1d2927aca0c3497cd36396fdde0a4cff41848"} Jan 26 09:37:17 crc kubenswrapper[4872]: I0126 09:37:17.496998 4872 scope.go:117] "RemoveContainer" containerID="0ccd9721504662de5f52f20d2d968d24d64df604f11270c047c06af15d559bc8" Jan 26 09:37:24 crc kubenswrapper[4872]: I0126 09:37:24.184181 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:37:24 crc kubenswrapper[4872]: E0126 09:37:24.184862 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:37:27 crc kubenswrapper[4872]: E0126 09:37:27.743688 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Jan 26 09:37:27 crc kubenswrapper[4872]: E0126 09:37:27.744449 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m5dt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-cd879_openshift-must-gather-6kdtd(2b6bd18e-9143-4c5d-b08b-9b497271d5a8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Jan 26 09:37:27 crc kubenswrapper[4872]: E0126 09:37:27.745562 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-6kdtd/crc-debug-cd879" podUID="2b6bd18e-9143-4c5d-b08b-9b497271d5a8" Jan 26 09:37:28 crc kubenswrapper[4872]: E0126 09:37:28.064364 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-6kdtd/crc-debug-cd879" podUID="2b6bd18e-9143-4c5d-b08b-9b497271d5a8" Jan 26 09:37:38 crc kubenswrapper[4872]: I0126 09:37:38.184165 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:37:38 crc kubenswrapper[4872]: E0126 09:37:38.184809 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:37:41 crc kubenswrapper[4872]: I0126 09:37:41.175286 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6kdtd/crc-debug-cd879" event={"ID":"2b6bd18e-9143-4c5d-b08b-9b497271d5a8","Type":"ContainerStarted","Data":"17307b3f5fd6b59cc70795888d7989a89426c403ce8803bd389d3603e3d78116"} Jan 26 09:37:41 crc kubenswrapper[4872]: I0126 09:37:41.194314 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-6kdtd/crc-debug-cd879" podStartSLOduration=1.69846018 podStartE2EDuration="30.194297779s" podCreationTimestamp="2026-01-26 09:37:11 +0000 UTC" firstStartedPulling="2026-01-26 09:37:12.274642484 +0000 UTC m=+1765.583482325" lastFinishedPulling="2026-01-26 09:37:40.770480123 +0000 UTC m=+1794.079319924" observedRunningTime="2026-01-26 09:37:41.192362442 +0000 UTC m=+1794.501202253" watchObservedRunningTime="2026-01-26 09:37:41.194297779 +0000 UTC m=+1794.503137580" Jan 26 09:37:52 crc kubenswrapper[4872]: I0126 09:37:52.184438 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:37:52 crc kubenswrapper[4872]: E0126 09:37:52.185477 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:38:06 crc kubenswrapper[4872]: I0126 09:38:06.183725 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:38:06 crc kubenswrapper[4872]: E0126 09:38:06.185026 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:38:16 crc kubenswrapper[4872]: E0126 09:38:16.682559 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b6bd18e_9143_4c5d_b08b_9b497271d5a8.slice/crio-17307b3f5fd6b59cc70795888d7989a89426c403ce8803bd389d3603e3d78116.scope\": RecentStats: unable to find data in memory cache]" Jan 26 09:38:17 crc kubenswrapper[4872]: I0126 09:38:17.503621 4872 generic.go:334] "Generic (PLEG): container finished" podID="2b6bd18e-9143-4c5d-b08b-9b497271d5a8" containerID="17307b3f5fd6b59cc70795888d7989a89426c403ce8803bd389d3603e3d78116" exitCode=0 Jan 26 09:38:17 crc kubenswrapper[4872]: I0126 09:38:17.503668 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6kdtd/crc-debug-cd879" event={"ID":"2b6bd18e-9143-4c5d-b08b-9b497271d5a8","Type":"ContainerDied","Data":"17307b3f5fd6b59cc70795888d7989a89426c403ce8803bd389d3603e3d78116"} Jan 26 09:38:18 crc kubenswrapper[4872]: I0126 09:38:18.609263 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/crc-debug-cd879" Jan 26 09:38:18 crc kubenswrapper[4872]: I0126 09:38:18.644721 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6kdtd/crc-debug-cd879"] Jan 26 09:38:18 crc kubenswrapper[4872]: I0126 09:38:18.653962 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6kdtd/crc-debug-cd879"] Jan 26 09:38:18 crc kubenswrapper[4872]: I0126 09:38:18.700240 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5dt7\" (UniqueName: \"kubernetes.io/projected/2b6bd18e-9143-4c5d-b08b-9b497271d5a8-kube-api-access-m5dt7\") pod \"2b6bd18e-9143-4c5d-b08b-9b497271d5a8\" (UID: \"2b6bd18e-9143-4c5d-b08b-9b497271d5a8\") " Jan 26 09:38:18 crc kubenswrapper[4872]: I0126 09:38:18.700287 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b6bd18e-9143-4c5d-b08b-9b497271d5a8-host\") pod \"2b6bd18e-9143-4c5d-b08b-9b497271d5a8\" (UID: \"2b6bd18e-9143-4c5d-b08b-9b497271d5a8\") " Jan 26 09:38:18 crc kubenswrapper[4872]: I0126 09:38:18.700579 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b6bd18e-9143-4c5d-b08b-9b497271d5a8-host" (OuterVolumeSpecName: "host") pod "2b6bd18e-9143-4c5d-b08b-9b497271d5a8" (UID: "2b6bd18e-9143-4c5d-b08b-9b497271d5a8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:38:18 crc kubenswrapper[4872]: I0126 09:38:18.706285 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b6bd18e-9143-4c5d-b08b-9b497271d5a8-kube-api-access-m5dt7" (OuterVolumeSpecName: "kube-api-access-m5dt7") pod "2b6bd18e-9143-4c5d-b08b-9b497271d5a8" (UID: "2b6bd18e-9143-4c5d-b08b-9b497271d5a8"). InnerVolumeSpecName "kube-api-access-m5dt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:38:18 crc kubenswrapper[4872]: I0126 09:38:18.802000 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5dt7\" (UniqueName: \"kubernetes.io/projected/2b6bd18e-9143-4c5d-b08b-9b497271d5a8-kube-api-access-m5dt7\") on node \"crc\" DevicePath \"\"" Jan 26 09:38:18 crc kubenswrapper[4872]: I0126 09:38:18.802039 4872 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2b6bd18e-9143-4c5d-b08b-9b497271d5a8-host\") on node \"crc\" DevicePath \"\"" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.196633 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b6bd18e-9143-4c5d-b08b-9b497271d5a8" path="/var/lib/kubelet/pods/2b6bd18e-9143-4c5d-b08b-9b497271d5a8/volumes" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.523332 4872 scope.go:117] "RemoveContainer" containerID="17307b3f5fd6b59cc70795888d7989a89426c403ce8803bd389d3603e3d78116" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.523378 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/crc-debug-cd879" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.801691 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6kdtd/crc-debug-2gkwk"] Jan 26 09:38:19 crc kubenswrapper[4872]: E0126 09:38:19.803259 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b6bd18e-9143-4c5d-b08b-9b497271d5a8" containerName="container-00" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.803345 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b6bd18e-9143-4c5d-b08b-9b497271d5a8" containerName="container-00" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.803614 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b6bd18e-9143-4c5d-b08b-9b497271d5a8" containerName="container-00" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.804298 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/crc-debug-2gkwk" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.806571 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-6kdtd"/"default-dockercfg-r2hgb" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.827023 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95gtp\" (UniqueName: \"kubernetes.io/projected/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5-kube-api-access-95gtp\") pod \"crc-debug-2gkwk\" (UID: \"3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5\") " pod="openshift-must-gather-6kdtd/crc-debug-2gkwk" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.827229 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5-host\") pod \"crc-debug-2gkwk\" (UID: \"3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5\") " pod="openshift-must-gather-6kdtd/crc-debug-2gkwk" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.929493 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5-host\") pod \"crc-debug-2gkwk\" (UID: \"3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5\") " pod="openshift-must-gather-6kdtd/crc-debug-2gkwk" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.929711 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95gtp\" (UniqueName: \"kubernetes.io/projected/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5-kube-api-access-95gtp\") pod \"crc-debug-2gkwk\" (UID: \"3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5\") " pod="openshift-must-gather-6kdtd/crc-debug-2gkwk" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.929760 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5-host\") pod \"crc-debug-2gkwk\" (UID: \"3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5\") " pod="openshift-must-gather-6kdtd/crc-debug-2gkwk" Jan 26 09:38:19 crc kubenswrapper[4872]: I0126 09:38:19.949742 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95gtp\" (UniqueName: \"kubernetes.io/projected/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5-kube-api-access-95gtp\") pod \"crc-debug-2gkwk\" (UID: \"3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5\") " pod="openshift-must-gather-6kdtd/crc-debug-2gkwk" Jan 26 09:38:20 crc kubenswrapper[4872]: I0126 09:38:20.118854 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/crc-debug-2gkwk" Jan 26 09:38:20 crc kubenswrapper[4872]: W0126 09:38:20.155279 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3cdbc8c9_f7a7_4f32_862a_8764d4ab27e5.slice/crio-5874d56fa9353b5ee7063b7e3ba2deb7e641da00e865bc1ebc99655049ae2f43 WatchSource:0}: Error finding container 5874d56fa9353b5ee7063b7e3ba2deb7e641da00e865bc1ebc99655049ae2f43: Status 404 returned error can't find the container with id 5874d56fa9353b5ee7063b7e3ba2deb7e641da00e865bc1ebc99655049ae2f43 Jan 26 09:38:20 crc kubenswrapper[4872]: I0126 09:38:20.183586 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:38:20 crc kubenswrapper[4872]: E0126 09:38:20.183875 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:38:20 crc kubenswrapper[4872]: I0126 09:38:20.536228 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6kdtd/crc-debug-2gkwk" event={"ID":"3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5","Type":"ContainerStarted","Data":"5874d56fa9353b5ee7063b7e3ba2deb7e641da00e865bc1ebc99655049ae2f43"} Jan 26 09:38:21 crc kubenswrapper[4872]: I0126 09:38:21.551220 4872 generic.go:334] "Generic (PLEG): container finished" podID="3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5" containerID="20188731c7464b77fbca1dad163d337c13e9f366fa74a3778ac5f255fdab77a7" exitCode=0 Jan 26 09:38:21 crc kubenswrapper[4872]: I0126 09:38:21.551306 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6kdtd/crc-debug-2gkwk" event={"ID":"3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5","Type":"ContainerDied","Data":"20188731c7464b77fbca1dad163d337c13e9f366fa74a3778ac5f255fdab77a7"} Jan 26 09:38:22 crc kubenswrapper[4872]: I0126 09:38:22.129258 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6kdtd/crc-debug-2gkwk"] Jan 26 09:38:22 crc kubenswrapper[4872]: I0126 09:38:22.136297 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6kdtd/crc-debug-2gkwk"] Jan 26 09:38:22 crc kubenswrapper[4872]: I0126 09:38:22.657878 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/crc-debug-2gkwk" Jan 26 09:38:22 crc kubenswrapper[4872]: I0126 09:38:22.685958 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95gtp\" (UniqueName: \"kubernetes.io/projected/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5-kube-api-access-95gtp\") pod \"3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5\" (UID: \"3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5\") " Jan 26 09:38:22 crc kubenswrapper[4872]: I0126 09:38:22.686447 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5-host\") pod \"3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5\" (UID: \"3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5\") " Jan 26 09:38:22 crc kubenswrapper[4872]: I0126 09:38:22.686521 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5-host" (OuterVolumeSpecName: "host") pod "3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5" (UID: "3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:38:22 crc kubenswrapper[4872]: I0126 09:38:22.687757 4872 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5-host\") on node \"crc\" DevicePath \"\"" Jan 26 09:38:22 crc kubenswrapper[4872]: I0126 09:38:22.691329 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5-kube-api-access-95gtp" (OuterVolumeSpecName: "kube-api-access-95gtp") pod "3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5" (UID: "3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5"). InnerVolumeSpecName "kube-api-access-95gtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:38:22 crc kubenswrapper[4872]: I0126 09:38:22.790678 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95gtp\" (UniqueName: \"kubernetes.io/projected/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5-kube-api-access-95gtp\") on node \"crc\" DevicePath \"\"" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.198414 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5" path="/var/lib/kubelet/pods/3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5/volumes" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.354541 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-6kdtd/crc-debug-n4wfm"] Jan 26 09:38:23 crc kubenswrapper[4872]: E0126 09:38:23.355011 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5" containerName="container-00" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.355035 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5" containerName="container-00" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.355328 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cdbc8c9-f7a7-4f32-862a-8764d4ab27e5" containerName="container-00" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.356205 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/crc-debug-n4wfm" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.404292 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d126c8a9-3775-448e-9fda-97c60de87d35-host\") pod \"crc-debug-n4wfm\" (UID: \"d126c8a9-3775-448e-9fda-97c60de87d35\") " pod="openshift-must-gather-6kdtd/crc-debug-n4wfm" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.404379 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmpw2\" (UniqueName: \"kubernetes.io/projected/d126c8a9-3775-448e-9fda-97c60de87d35-kube-api-access-bmpw2\") pod \"crc-debug-n4wfm\" (UID: \"d126c8a9-3775-448e-9fda-97c60de87d35\") " pod="openshift-must-gather-6kdtd/crc-debug-n4wfm" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.506707 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d126c8a9-3775-448e-9fda-97c60de87d35-host\") pod \"crc-debug-n4wfm\" (UID: \"d126c8a9-3775-448e-9fda-97c60de87d35\") " pod="openshift-must-gather-6kdtd/crc-debug-n4wfm" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.506760 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmpw2\" (UniqueName: \"kubernetes.io/projected/d126c8a9-3775-448e-9fda-97c60de87d35-kube-api-access-bmpw2\") pod \"crc-debug-n4wfm\" (UID: \"d126c8a9-3775-448e-9fda-97c60de87d35\") " pod="openshift-must-gather-6kdtd/crc-debug-n4wfm" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.507117 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d126c8a9-3775-448e-9fda-97c60de87d35-host\") pod \"crc-debug-n4wfm\" (UID: \"d126c8a9-3775-448e-9fda-97c60de87d35\") " pod="openshift-must-gather-6kdtd/crc-debug-n4wfm" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.542283 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmpw2\" (UniqueName: \"kubernetes.io/projected/d126c8a9-3775-448e-9fda-97c60de87d35-kube-api-access-bmpw2\") pod \"crc-debug-n4wfm\" (UID: \"d126c8a9-3775-448e-9fda-97c60de87d35\") " pod="openshift-must-gather-6kdtd/crc-debug-n4wfm" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.572310 4872 scope.go:117] "RemoveContainer" containerID="20188731c7464b77fbca1dad163d337c13e9f366fa74a3778ac5f255fdab77a7" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.572373 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/crc-debug-2gkwk" Jan 26 09:38:23 crc kubenswrapper[4872]: I0126 09:38:23.675863 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/crc-debug-n4wfm" Jan 26 09:38:24 crc kubenswrapper[4872]: I0126 09:38:24.583822 4872 generic.go:334] "Generic (PLEG): container finished" podID="d126c8a9-3775-448e-9fda-97c60de87d35" containerID="08c6c4b9b7389d76ecffa62d41ea4673a7136ab2952d4bfb9c965a94d5861d11" exitCode=0 Jan 26 09:38:24 crc kubenswrapper[4872]: I0126 09:38:24.583863 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6kdtd/crc-debug-n4wfm" event={"ID":"d126c8a9-3775-448e-9fda-97c60de87d35","Type":"ContainerDied","Data":"08c6c4b9b7389d76ecffa62d41ea4673a7136ab2952d4bfb9c965a94d5861d11"} Jan 26 09:38:24 crc kubenswrapper[4872]: I0126 09:38:24.584201 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6kdtd/crc-debug-n4wfm" event={"ID":"d126c8a9-3775-448e-9fda-97c60de87d35","Type":"ContainerStarted","Data":"f28a92cc0fbb0b16d1583006ca93070df5ed13c57066fed92182f102eb0a455e"} Jan 26 09:38:24 crc kubenswrapper[4872]: I0126 09:38:24.623526 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6kdtd/crc-debug-n4wfm"] Jan 26 09:38:24 crc kubenswrapper[4872]: I0126 09:38:24.636484 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6kdtd/crc-debug-n4wfm"] Jan 26 09:38:25 crc kubenswrapper[4872]: I0126 09:38:25.695374 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/crc-debug-n4wfm" Jan 26 09:38:25 crc kubenswrapper[4872]: I0126 09:38:25.749777 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmpw2\" (UniqueName: \"kubernetes.io/projected/d126c8a9-3775-448e-9fda-97c60de87d35-kube-api-access-bmpw2\") pod \"d126c8a9-3775-448e-9fda-97c60de87d35\" (UID: \"d126c8a9-3775-448e-9fda-97c60de87d35\") " Jan 26 09:38:25 crc kubenswrapper[4872]: I0126 09:38:25.749909 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d126c8a9-3775-448e-9fda-97c60de87d35-host\") pod \"d126c8a9-3775-448e-9fda-97c60de87d35\" (UID: \"d126c8a9-3775-448e-9fda-97c60de87d35\") " Jan 26 09:38:25 crc kubenswrapper[4872]: I0126 09:38:25.750028 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d126c8a9-3775-448e-9fda-97c60de87d35-host" (OuterVolumeSpecName: "host") pod "d126c8a9-3775-448e-9fda-97c60de87d35" (UID: "d126c8a9-3775-448e-9fda-97c60de87d35"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 26 09:38:25 crc kubenswrapper[4872]: I0126 09:38:25.750627 4872 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d126c8a9-3775-448e-9fda-97c60de87d35-host\") on node \"crc\" DevicePath \"\"" Jan 26 09:38:25 crc kubenswrapper[4872]: I0126 09:38:25.755175 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d126c8a9-3775-448e-9fda-97c60de87d35-kube-api-access-bmpw2" (OuterVolumeSpecName: "kube-api-access-bmpw2") pod "d126c8a9-3775-448e-9fda-97c60de87d35" (UID: "d126c8a9-3775-448e-9fda-97c60de87d35"). InnerVolumeSpecName "kube-api-access-bmpw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:38:25 crc kubenswrapper[4872]: I0126 09:38:25.852964 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmpw2\" (UniqueName: \"kubernetes.io/projected/d126c8a9-3775-448e-9fda-97c60de87d35-kube-api-access-bmpw2\") on node \"crc\" DevicePath \"\"" Jan 26 09:38:26 crc kubenswrapper[4872]: I0126 09:38:26.600551 4872 scope.go:117] "RemoveContainer" containerID="08c6c4b9b7389d76ecffa62d41ea4673a7136ab2952d4bfb9c965a94d5861d11" Jan 26 09:38:26 crc kubenswrapper[4872]: I0126 09:38:26.600602 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/crc-debug-n4wfm" Jan 26 09:38:27 crc kubenswrapper[4872]: I0126 09:38:27.193475 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d126c8a9-3775-448e-9fda-97c60de87d35" path="/var/lib/kubelet/pods/d126c8a9-3775-448e-9fda-97c60de87d35/volumes" Jan 26 09:38:31 crc kubenswrapper[4872]: I0126 09:38:31.183870 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:38:31 crc kubenswrapper[4872]: E0126 09:38:31.185926 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:38:42 crc kubenswrapper[4872]: I0126 09:38:42.307827 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-75b468fcb8-mdk2p_b0f0f467-e8ad-4766-86e5-b46271a39b39/barbican-api/0.log" Jan 26 09:38:42 crc kubenswrapper[4872]: I0126 09:38:42.464854 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-75b468fcb8-mdk2p_b0f0f467-e8ad-4766-86e5-b46271a39b39/barbican-api-log/0.log" Jan 26 09:38:42 crc kubenswrapper[4872]: I0126 09:38:42.525336 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-664886958b-t96tw_6532ea83-7260-4883-8cdd-236c90c5c90a/barbican-keystone-listener/0.log" Jan 26 09:38:42 crc kubenswrapper[4872]: I0126 09:38:42.635200 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-664886958b-t96tw_6532ea83-7260-4883-8cdd-236c90c5c90a/barbican-keystone-listener-log/0.log" Jan 26 09:38:42 crc kubenswrapper[4872]: I0126 09:38:42.723736 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5f44467695-j8db6_1dc34b32-f172-4b29-8bd1-a9f7a86d0996/barbican-worker/0.log" Jan 26 09:38:42 crc kubenswrapper[4872]: I0126 09:38:42.763684 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5f44467695-j8db6_1dc34b32-f172-4b29-8bd1-a9f7a86d0996/barbican-worker-log/0.log" Jan 26 09:38:42 crc kubenswrapper[4872]: I0126 09:38:42.938649 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f535322c-098d-4f83-b91b-8cd72dd3339b/ceilometer-central-agent/0.log" Jan 26 09:38:42 crc kubenswrapper[4872]: I0126 09:38:42.994647 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f535322c-098d-4f83-b91b-8cd72dd3339b/proxy-httpd/0.log" Jan 26 09:38:43 crc kubenswrapper[4872]: I0126 09:38:43.008006 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f535322c-098d-4f83-b91b-8cd72dd3339b/ceilometer-notification-agent/0.log" Jan 26 09:38:43 crc kubenswrapper[4872]: I0126 09:38:43.043619 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_f535322c-098d-4f83-b91b-8cd72dd3339b/sg-core/0.log" Jan 26 09:38:43 crc kubenswrapper[4872]: I0126 09:38:43.190288 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d9c67aaa-15db-4359-b978-685a7ed15ec4/cinder-api-log/0.log" Jan 26 09:38:43 crc kubenswrapper[4872]: I0126 09:38:43.212387 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_d9c67aaa-15db-4359-b978-685a7ed15ec4/cinder-api/0.log" Jan 26 09:38:43 crc kubenswrapper[4872]: I0126 09:38:43.365887 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6faf686b-43c3-44f2-9e84-a142ce8192d1/cinder-scheduler/0.log" Jan 26 09:38:43 crc kubenswrapper[4872]: I0126 09:38:43.394312 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6faf686b-43c3-44f2-9e84-a142ce8192d1/cinder-scheduler/1.log" Jan 26 09:38:43 crc kubenswrapper[4872]: I0126 09:38:43.425760 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_6faf686b-43c3-44f2-9e84-a142ce8192d1/probe/0.log" Jan 26 09:38:43 crc kubenswrapper[4872]: I0126 09:38:43.558635 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-59cf4bdb65-nzwnv_4654462c-cb3b-46d4-aec6-9e701038aebd/init/0.log" Jan 26 09:38:43 crc kubenswrapper[4872]: I0126 09:38:43.926156 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-59cf4bdb65-nzwnv_4654462c-cb3b-46d4-aec6-9e701038aebd/init/0.log" Jan 26 09:38:43 crc kubenswrapper[4872]: I0126 09:38:43.929102 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-59cf4bdb65-nzwnv_4654462c-cb3b-46d4-aec6-9e701038aebd/dnsmasq-dns/0.log" Jan 26 09:38:43 crc kubenswrapper[4872]: I0126 09:38:43.935753 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4e79d158-7664-4bfd-98b8-8cbf9092584a/glance-httpd/0.log" Jan 26 09:38:44 crc kubenswrapper[4872]: I0126 09:38:44.107386 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_4e79d158-7664-4bfd-98b8-8cbf9092584a/glance-log/0.log" Jan 26 09:38:44 crc kubenswrapper[4872]: I0126 09:38:44.162520 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6/glance-httpd/0.log" Jan 26 09:38:44 crc kubenswrapper[4872]: I0126 09:38:44.181379 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_e7c8b404-f2e3-4d1e-a5ac-cd6651d48bf6/glance-log/0.log" Jan 26 09:38:44 crc kubenswrapper[4872]: I0126 09:38:44.354773 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-6785557d58-cspfm_30773720-c572-485d-8ea3-3ad3293e4869/init/0.log" Jan 26 09:38:44 crc kubenswrapper[4872]: I0126 09:38:44.515922 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-6785557d58-cspfm_30773720-c572-485d-8ea3-3ad3293e4869/ironic-api-log/0.log" Jan 26 09:38:44 crc kubenswrapper[4872]: I0126 09:38:44.583886 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-6785557d58-cspfm_30773720-c572-485d-8ea3-3ad3293e4869/ironic-api/0.log" Jan 26 09:38:44 crc kubenswrapper[4872]: I0126 09:38:44.645302 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-6785557d58-cspfm_30773720-c572-485d-8ea3-3ad3293e4869/init/0.log" Jan 26 09:38:44 crc kubenswrapper[4872]: I0126 09:38:44.747373 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/init/0.log" Jan 26 09:38:44 crc kubenswrapper[4872]: I0126 09:38:44.936085 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/init/0.log" Jan 26 09:38:45 crc kubenswrapper[4872]: I0126 09:38:45.021188 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/ironic-python-agent-init/0.log" Jan 26 09:38:45 crc kubenswrapper[4872]: I0126 09:38:45.062646 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/ironic-python-agent-init/0.log" Jan 26 09:38:45 crc kubenswrapper[4872]: I0126 09:38:45.282330 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/ironic-python-agent-init/0.log" Jan 26 09:38:45 crc kubenswrapper[4872]: I0126 09:38:45.316029 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/init/0.log" Jan 26 09:38:45 crc kubenswrapper[4872]: I0126 09:38:45.674894 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/init/0.log" Jan 26 09:38:45 crc kubenswrapper[4872]: I0126 09:38:45.694903 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/pxe-init/0.log" Jan 26 09:38:45 crc kubenswrapper[4872]: I0126 09:38:45.971320 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/ironic-python-agent-init/0.log" Jan 26 09:38:46 crc kubenswrapper[4872]: I0126 09:38:46.150681 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/httpboot/0.log" Jan 26 09:38:46 crc kubenswrapper[4872]: I0126 09:38:46.183549 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:38:46 crc kubenswrapper[4872]: E0126 09:38:46.183809 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:38:46 crc kubenswrapper[4872]: I0126 09:38:46.302104 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/ironic-conductor/0.log" Jan 26 09:38:46 crc kubenswrapper[4872]: I0126 09:38:46.367202 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/ramdisk-logs/0.log" Jan 26 09:38:46 crc kubenswrapper[4872]: I0126 09:38:46.543680 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-8dnjd_3e385f5e-2610-44cd-ac51-ff0642cb47a9/init/0.log" Jan 26 09:38:46 crc kubenswrapper[4872]: I0126 09:38:46.605465 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/pxe-init/0.log" Jan 26 09:38:46 crc kubenswrapper[4872]: I0126 09:38:46.694793 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/pxe-init/0.log" Jan 26 09:38:46 crc kubenswrapper[4872]: I0126 09:38:46.803562 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-8dnjd_3e385f5e-2610-44cd-ac51-ff0642cb47a9/init/0.log" Jan 26 09:38:46 crc kubenswrapper[4872]: I0126 09:38:46.908191 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-8dnjd_3e385f5e-2610-44cd-ac51-ff0642cb47a9/ironic-db-sync/0.log" Jan 26 09:38:46 crc kubenswrapper[4872]: I0126 09:38:46.938638 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_e135958c-61c2-4199-94b3-2a8e6623310c/pxe-init/0.log" Jan 26 09:38:46 crc kubenswrapper[4872]: I0126 09:38:46.940640 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_8865ee9d-9129-488e-84e2-cf1a11547f8a/ironic-python-agent-init/0.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.094403 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_8865ee9d-9129-488e-84e2-cf1a11547f8a/ironic-python-agent-init/0.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.116415 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_8865ee9d-9129-488e-84e2-cf1a11547f8a/inspector-pxe-init/0.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.153360 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_8865ee9d-9129-488e-84e2-cf1a11547f8a/inspector-pxe-init/0.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.316684 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_8865ee9d-9129-488e-84e2-cf1a11547f8a/ironic-python-agent-init/0.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.327220 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_8865ee9d-9129-488e-84e2-cf1a11547f8a/inspector-pxe-init/0.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.372758 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_8865ee9d-9129-488e-84e2-cf1a11547f8a/ironic-inspector/1.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.452914 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_8865ee9d-9129-488e-84e2-cf1a11547f8a/inspector-httpboot/0.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.494129 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_8865ee9d-9129-488e-84e2-cf1a11547f8a/ironic-inspector/2.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.514183 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_8865ee9d-9129-488e-84e2-cf1a11547f8a/ramdisk-logs/0.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.557063 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_8865ee9d-9129-488e-84e2-cf1a11547f8a/ironic-inspector-httpd/0.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.858708 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-db-sync-dv876_909a12a7-4b22-43b8-8d96-1fff1ef2738c/ironic-inspector-db-sync/0.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.861490 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-neutron-agent-5b47d6fc7b-7wkw2_f721aeee-7daf-413d-885b-5c146881eb99/ironic-neutron-agent/3.log" Jan 26 09:38:47 crc kubenswrapper[4872]: I0126 09:38:47.888752 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-neutron-agent-5b47d6fc7b-7wkw2_f721aeee-7daf-413d-885b-5c146881eb99/ironic-neutron-agent/2.log" Jan 26 09:38:48 crc kubenswrapper[4872]: I0126 09:38:48.061881 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_2917f9e7-80a7-4af0-9816-aa58a9684f94/kube-state-metrics/0.log" Jan 26 09:38:48 crc kubenswrapper[4872]: I0126 09:38:48.189982 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-769b85d4b6-8zpg4_5f3edc0f-a2a4-42ee-8f69-44195674cb06/keystone-api/0.log" Jan 26 09:38:48 crc kubenswrapper[4872]: I0126 09:38:48.458118 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-795668bf-dk6bd_a0396e21-0a34-49b9-aa03-64c838f0da0c/neutron-httpd/0.log" Jan 26 09:38:48 crc kubenswrapper[4872]: I0126 09:38:48.555711 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-795668bf-dk6bd_a0396e21-0a34-49b9-aa03-64c838f0da0c/neutron-api/0.log" Jan 26 09:38:48 crc kubenswrapper[4872]: I0126 09:38:48.668394 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_12b5f34c-5cc6-448f-ac2f-420df97eaaa3/nova-api-api/0.log" Jan 26 09:38:48 crc kubenswrapper[4872]: I0126 09:38:48.729821 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_12b5f34c-5cc6-448f-ac2f-420df97eaaa3/nova-api-log/0.log" Jan 26 09:38:48 crc kubenswrapper[4872]: I0126 09:38:48.862497 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_d34fd0f4-3e8c-42a8-a242-5c0bea758f42/nova-cell0-conductor-conductor/0.log" Jan 26 09:38:48 crc kubenswrapper[4872]: I0126 09:38:48.968375 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_dc5d286f-e785-4a4b-b96e-392ea79d7d2e/nova-cell1-conductor-conductor/0.log" Jan 26 09:38:49 crc kubenswrapper[4872]: I0126 09:38:49.169848 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8903a57e-a76b-42a5-a025-992c12eab47c/nova-cell1-novncproxy-novncproxy/0.log" Jan 26 09:38:49 crc kubenswrapper[4872]: I0126 09:38:49.314712 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_735c976b-3cf3-4b78-b5fc-4d92d09f410c/nova-metadata-log/0.log" Jan 26 09:38:49 crc kubenswrapper[4872]: I0126 09:38:49.492838 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_c97bb470-8700-44dc-99a4-4e39b14aa332/nova-scheduler-scheduler/0.log" Jan 26 09:38:49 crc kubenswrapper[4872]: I0126 09:38:49.577837 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ab5ce366-26e9-424a-9a73-9173636c2902/mysql-bootstrap/0.log" Jan 26 09:38:49 crc kubenswrapper[4872]: I0126 09:38:49.712251 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_735c976b-3cf3-4b78-b5fc-4d92d09f410c/nova-metadata-metadata/0.log" Jan 26 09:38:49 crc kubenswrapper[4872]: I0126 09:38:49.747269 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ab5ce366-26e9-424a-9a73-9173636c2902/mysql-bootstrap/0.log" Jan 26 09:38:49 crc kubenswrapper[4872]: I0126 09:38:49.893224 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_ab5ce366-26e9-424a-9a73-9173636c2902/galera/0.log" Jan 26 09:38:49 crc kubenswrapper[4872]: I0126 09:38:49.937805 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9082961a-1e93-49b5-8420-7516b675ea6f/mysql-bootstrap/0.log" Jan 26 09:38:50 crc kubenswrapper[4872]: I0126 09:38:50.121083 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_52a92868-e230-4403-9297-51c2b457343b/openstackclient/0.log" Jan 26 09:38:50 crc kubenswrapper[4872]: I0126 09:38:50.150623 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9082961a-1e93-49b5-8420-7516b675ea6f/galera/0.log" Jan 26 09:38:50 crc kubenswrapper[4872]: I0126 09:38:50.174312 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9082961a-1e93-49b5-8420-7516b675ea6f/mysql-bootstrap/0.log" Jan 26 09:38:50 crc kubenswrapper[4872]: I0126 09:38:50.344753 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-gmp2c_cb72fdef-372c-41a1-bf57-6c71162ab194/ovn-controller/0.log" Jan 26 09:38:50 crc kubenswrapper[4872]: I0126 09:38:50.385360 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-5x7dg_2cdaf461-41f0-44d3-8fad-d852c276c343/openstack-network-exporter/0.log" Jan 26 09:38:50 crc kubenswrapper[4872]: I0126 09:38:50.554343 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mgn5r_4ece36ce-cbb6-4012-8b5c-87834502f26b/ovsdb-server-init/0.log" Jan 26 09:38:50 crc kubenswrapper[4872]: I0126 09:38:50.994548 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mgn5r_4ece36ce-cbb6-4012-8b5c-87834502f26b/ovs-vswitchd/0.log" Jan 26 09:38:51 crc kubenswrapper[4872]: I0126 09:38:51.019112 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mgn5r_4ece36ce-cbb6-4012-8b5c-87834502f26b/ovsdb-server/0.log" Jan 26 09:38:51 crc kubenswrapper[4872]: I0126 09:38:51.020685 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mgn5r_4ece36ce-cbb6-4012-8b5c-87834502f26b/ovsdb-server-init/0.log" Jan 26 09:38:51 crc kubenswrapper[4872]: I0126 09:38:51.282070 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_fc8642bd-5341-4e95-be61-deb51d754edb/openstack-network-exporter/0.log" Jan 26 09:38:51 crc kubenswrapper[4872]: I0126 09:38:51.285423 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_fc8642bd-5341-4e95-be61-deb51d754edb/ovn-northd/0.log" Jan 26 09:38:51 crc kubenswrapper[4872]: I0126 09:38:51.342408 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bf0544f2-34dc-45fc-986e-cb2554c04566/openstack-network-exporter/0.log" Jan 26 09:38:51 crc kubenswrapper[4872]: I0126 09:38:51.527239 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_204a2939-a385-4f7d-a08b-a008e6a14df0/openstack-network-exporter/0.log" Jan 26 09:38:51 crc kubenswrapper[4872]: I0126 09:38:51.551823 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_bf0544f2-34dc-45fc-986e-cb2554c04566/ovsdbserver-nb/0.log" Jan 26 09:38:51 crc kubenswrapper[4872]: I0126 09:38:51.611691 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_204a2939-a385-4f7d-a08b-a008e6a14df0/ovsdbserver-sb/0.log" Jan 26 09:38:51 crc kubenswrapper[4872]: I0126 09:38:51.713119 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-b65567844-p9ckk_e643ca60-8ad6-4846-976d-4d2f57c70788/placement-api/0.log" Jan 26 09:38:51 crc kubenswrapper[4872]: I0126 09:38:51.830585 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-b65567844-p9ckk_e643ca60-8ad6-4846-976d-4d2f57c70788/placement-log/0.log" Jan 26 09:38:51 crc kubenswrapper[4872]: I0126 09:38:51.959203 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3900451f-25bd-4b04-b516-20ff2a5c8233/setup-container/0.log" Jan 26 09:38:52 crc kubenswrapper[4872]: I0126 09:38:52.171137 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fac8c53c-96c3-491f-9555-16a797c45699/setup-container/0.log" Jan 26 09:38:52 crc kubenswrapper[4872]: I0126 09:38:52.193123 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3900451f-25bd-4b04-b516-20ff2a5c8233/setup-container/0.log" Jan 26 09:38:52 crc kubenswrapper[4872]: I0126 09:38:52.222375 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_3900451f-25bd-4b04-b516-20ff2a5c8233/rabbitmq/0.log" Jan 26 09:38:52 crc kubenswrapper[4872]: I0126 09:38:52.422703 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fac8c53c-96c3-491f-9555-16a797c45699/rabbitmq/0.log" Jan 26 09:38:52 crc kubenswrapper[4872]: I0126 09:38:52.454290 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fac8c53c-96c3-491f-9555-16a797c45699/setup-container/0.log" Jan 26 09:38:52 crc kubenswrapper[4872]: I0126 09:38:52.524995 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-68bdb4f9bc-lg4kg_c15c32f5-7a02-497d-a689-f3c82c63ad8b/proxy-httpd/0.log" Jan 26 09:38:52 crc kubenswrapper[4872]: I0126 09:38:52.661574 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-68bdb4f9bc-lg4kg_c15c32f5-7a02-497d-a689-f3c82c63ad8b/proxy-server/0.log" Jan 26 09:38:52 crc kubenswrapper[4872]: I0126 09:38:52.733333 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-nvsn5_3ffefe20-1e04-4a87-b2cf-5140aefbb87d/swift-ring-rebalance/0.log" Jan 26 09:38:52 crc kubenswrapper[4872]: I0126 09:38:52.883342 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/account-auditor/0.log" Jan 26 09:38:52 crc kubenswrapper[4872]: I0126 09:38:52.928424 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/account-reaper/0.log" Jan 26 09:38:52 crc kubenswrapper[4872]: I0126 09:38:52.974232 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/account-replicator/0.log" Jan 26 09:38:53 crc kubenswrapper[4872]: I0126 09:38:53.047948 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/container-auditor/0.log" Jan 26 09:38:53 crc kubenswrapper[4872]: I0126 09:38:53.123504 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/account-server/0.log" Jan 26 09:38:53 crc kubenswrapper[4872]: I0126 09:38:53.145254 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/container-replicator/0.log" Jan 26 09:38:53 crc kubenswrapper[4872]: I0126 09:38:53.231855 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/container-server/0.log" Jan 26 09:38:53 crc kubenswrapper[4872]: I0126 09:38:53.267373 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/container-updater/0.log" Jan 26 09:38:53 crc kubenswrapper[4872]: I0126 09:38:53.381322 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/object-expirer/0.log" Jan 26 09:38:53 crc kubenswrapper[4872]: I0126 09:38:53.401284 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/object-auditor/0.log" Jan 26 09:38:53 crc kubenswrapper[4872]: I0126 09:38:53.497940 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/object-replicator/0.log" Jan 26 09:38:53 crc kubenswrapper[4872]: I0126 09:38:53.539260 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/object-server/0.log" Jan 26 09:38:53 crc kubenswrapper[4872]: I0126 09:38:53.636419 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/rsync/0.log" Jan 26 09:38:53 crc kubenswrapper[4872]: I0126 09:38:53.638974 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/object-updater/0.log" Jan 26 09:38:53 crc kubenswrapper[4872]: I0126 09:38:53.743099 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_f94cf23a-424d-4bd6-8e19-b1f3dc7a0160/swift-recon-cron/0.log" Jan 26 09:38:55 crc kubenswrapper[4872]: I0126 09:38:55.778120 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_98a9098c-4abc-4a20-be26-d62bd842bd04/memcached/0.log" Jan 26 09:38:58 crc kubenswrapper[4872]: I0126 09:38:58.184191 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:38:58 crc kubenswrapper[4872]: E0126 09:38:58.184481 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:39:10 crc kubenswrapper[4872]: I0126 09:39:10.184296 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:39:10 crc kubenswrapper[4872]: E0126 09:39:10.185175 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:39:17 crc kubenswrapper[4872]: I0126 09:39:17.449747 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7f86f8796f-96757_866888fd-6bb1-4184-9e17-352cb8d3727f/manager/0.log" Jan 26 09:39:17 crc kubenswrapper[4872]: I0126 09:39:17.604585 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz_c19dba91-7b40-4824-9377-1e55368da0c7/util/0.log" Jan 26 09:39:17 crc kubenswrapper[4872]: I0126 09:39:17.809945 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz_c19dba91-7b40-4824-9377-1e55368da0c7/pull/0.log" Jan 26 09:39:17 crc kubenswrapper[4872]: I0126 09:39:17.813700 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz_c19dba91-7b40-4824-9377-1e55368da0c7/util/0.log" Jan 26 09:39:17 crc kubenswrapper[4872]: I0126 09:39:17.820323 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz_c19dba91-7b40-4824-9377-1e55368da0c7/pull/0.log" Jan 26 09:39:18 crc kubenswrapper[4872]: I0126 09:39:18.006002 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz_c19dba91-7b40-4824-9377-1e55368da0c7/pull/0.log" Jan 26 09:39:18 crc kubenswrapper[4872]: I0126 09:39:18.006509 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz_c19dba91-7b40-4824-9377-1e55368da0c7/extract/0.log" Jan 26 09:39:18 crc kubenswrapper[4872]: I0126 09:39:18.016763 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bb7847681d6a79649e7297ea5ea7118c9478f5558bd16c37862aaf7d90dsbpz_c19dba91-7b40-4824-9377-1e55368da0c7/util/0.log" Jan 26 09:39:18 crc kubenswrapper[4872]: I0126 09:39:18.230693 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7478f7dbf9-cvcnq_b3751f8d-5921-4b5f-b740-1ba5fc532f7b/manager/0.log" Jan 26 09:39:18 crc kubenswrapper[4872]: I0126 09:39:18.266091 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-b45d7bf98-ldqqh_a313f841-7d48-4621-a385-b001d8cfa11b/manager/0.log" Jan 26 09:39:18 crc kubenswrapper[4872]: I0126 09:39:18.463107 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-594c8c9d5d-g5txh_c5bfadfb-1ac5-475e-8b5f-8096dd00d850/manager/0.log" Jan 26 09:39:18 crc kubenswrapper[4872]: I0126 09:39:18.478995 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78fdd796fd-cwkcb_7f35eab0-a711-4b5d-8b5a-05c72414c30c/manager/0.log" Jan 26 09:39:18 crc kubenswrapper[4872]: I0126 09:39:18.645940 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-77d5c5b54f-s4b5m_2bda61c9-b3dd-476a-86fb-882a5f172ba9/manager/0.log" Jan 26 09:39:18 crc kubenswrapper[4872]: I0126 09:39:18.888209 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-758868c854-rss8g_5619ad3b-5066-4163-9cff-7654030fc5b1/manager/0.log" Jan 26 09:39:18 crc kubenswrapper[4872]: I0126 09:39:18.902274 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6b46765b57-crcq2_9cdfe7c6-5b86-4204-b9bf-3597361af9fe/manager/0.log" Jan 26 09:39:19 crc kubenswrapper[4872]: I0126 09:39:19.023545 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b8b6d4659-q4kjt_87e00bbc-66f4-4921-94ce-24d7b658341c/manager/0.log" Jan 26 09:39:19 crc kubenswrapper[4872]: I0126 09:39:19.099244 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-78c6999f6f-wct5s_4d3db7ee-d472-427c-96a2-1c0f67a3b507/manager/0.log" Jan 26 09:39:19 crc kubenswrapper[4872]: I0126 09:39:19.268110 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6b9fb5fdcb-tcc58_40ff0d90-ae83-47b6-b6b0-8969b47bcf92/manager/0.log" Jan 26 09:39:19 crc kubenswrapper[4872]: I0126 09:39:19.360876 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-78d58447c5-qwp9z_c05670e5-8312-46ce-9fe7-da7ecc691465/manager/0.log" Jan 26 09:39:19 crc kubenswrapper[4872]: I0126 09:39:19.502548 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-7bdb645866-vz57m_acf08f03-b388-4e2a-99a2-2d492e490e67/manager/0.log" Jan 26 09:39:19 crc kubenswrapper[4872]: I0126 09:39:19.512576 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4cd88d46-77t7p_148da344-f5f5-43ab-ba3c-342f6cc10b47/manager/0.log" Jan 26 09:39:19 crc kubenswrapper[4872]: I0126 09:39:19.703586 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6b68b8b854fjdzv_a0993a22-0883-4db0-9c07-c30657d9fb37/manager/0.log" Jan 26 09:39:19 crc kubenswrapper[4872]: I0126 09:39:19.816914 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-8578c7d94c-gfq28_8582c505-0f00-4751-bae0-b9324821e2cb/operator/0.log" Jan 26 09:39:20 crc kubenswrapper[4872]: I0126 09:39:20.099708 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-psjm6_1b7d4719-ae54-402a-83c2-39ca5edb30d0/registry-server/0.log" Jan 26 09:39:20 crc kubenswrapper[4872]: I0126 09:39:20.210752 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f75f45d54-r7rv2_37007861-ed2f-4453-8a25-aa39f6b0f347/manager/0.log" Jan 26 09:39:20 crc kubenswrapper[4872]: I0126 09:39:20.308771 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-79d5ccc684-7fm49_60fd216c-7ecc-477f-ac29-b7334ea5e450/manager/0.log" Jan 26 09:39:20 crc kubenswrapper[4872]: I0126 09:39:20.594029 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-69bc8c84fd-8f9kj_6e3c0ac4-80fa-4d8d-864d-38056dc22402/manager/0.log" Jan 26 09:39:20 crc kubenswrapper[4872]: I0126 09:39:20.595022 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-hmfxd_1fa3edb5-2354-4c32-ac85-d6be3a7dc16c/operator/0.log" Jan 26 09:39:20 crc kubenswrapper[4872]: I0126 09:39:20.680414 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-547cbdb99f-cqr4g_d92c1258-d114-4696-a481-f9fff7faa046/manager/0.log" Jan 26 09:39:20 crc kubenswrapper[4872]: I0126 09:39:20.828881 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-85cd9769bb-zfvrw_4b97d503-908a-48b0-bb0c-5e2cdcf251af/manager/0.log" Jan 26 09:39:20 crc kubenswrapper[4872]: I0126 09:39:20.880148 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-69797bbcbd-tdnns_12d27df7-ee8b-4e48-b0fe-0b1f018e71f8/manager/0.log" Jan 26 09:39:21 crc kubenswrapper[4872]: I0126 09:39:21.025681 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-564965969-hfj2p_ced935ab-f3e5-470d-aa18-e15bb944fb86/manager/0.log" Jan 26 09:39:21 crc kubenswrapper[4872]: I0126 09:39:21.185163 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:39:21 crc kubenswrapper[4872]: E0126 09:39:21.185928 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:39:32 crc kubenswrapper[4872]: I0126 09:39:32.184095 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:39:32 crc kubenswrapper[4872]: E0126 09:39:32.185274 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:39:39 crc kubenswrapper[4872]: I0126 09:39:39.521104 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-hvnl5_9b1a66ff-c739-4077-ab76-bf675e158a77/control-plane-machine-set-operator/0.log" Jan 26 09:39:39 crc kubenswrapper[4872]: I0126 09:39:39.700830 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-x9zlw_07ffe425-ff10-4c9f-b3c5-fc2c5e167f67/kube-rbac-proxy/0.log" Jan 26 09:39:39 crc kubenswrapper[4872]: I0126 09:39:39.759159 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-x9zlw_07ffe425-ff10-4c9f-b3c5-fc2c5e167f67/machine-api-operator/0.log" Jan 26 09:39:45 crc kubenswrapper[4872]: I0126 09:39:45.184440 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:39:46 crc kubenswrapper[4872]: I0126 09:39:46.319886 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerStarted","Data":"8dad67da6bdcbaf03f302de0fdcef76cd940f73a0eab7a62b9aac33c1f3c41aa"} Jan 26 09:39:53 crc kubenswrapper[4872]: I0126 09:39:53.209532 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-m97zb_1d91c9be-4442-4976-8714-bf07b8777576/cert-manager-controller/0.log" Jan 26 09:39:53 crc kubenswrapper[4872]: I0126 09:39:53.412895 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-jnspw_f1d545e1-0b8c-481e-a979-760d92eec6da/cert-manager-cainjector/0.log" Jan 26 09:39:53 crc kubenswrapper[4872]: I0126 09:39:53.448589 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-hxz6l_e165986a-e1a8-47bc-8e3c-2609509f1ede/cert-manager-webhook/0.log" Jan 26 09:40:06 crc kubenswrapper[4872]: I0126 09:40:06.268076 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7754f76f8b-lk6cm_fead2890-af05-404f-a7d4-6321b48e31a2/nmstate-console-plugin/0.log" Jan 26 09:40:06 crc kubenswrapper[4872]: I0126 09:40:06.416952 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-cbbnl_d7ecce6c-87f7-4b85-aceb-1430b1668e0d/nmstate-handler/0.log" Jan 26 09:40:06 crc kubenswrapper[4872]: I0126 09:40:06.441638 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-rzhxm_464e800f-b945-4d14-8fb6-d023d82ac993/kube-rbac-proxy/0.log" Jan 26 09:40:06 crc kubenswrapper[4872]: I0126 09:40:06.475423 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-54757c584b-rzhxm_464e800f-b945-4d14-8fb6-d023d82ac993/nmstate-metrics/0.log" Jan 26 09:40:06 crc kubenswrapper[4872]: I0126 09:40:06.679206 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-646758c888-b72fb_d45d048b-dc90-4af0-800d-cdaa152c7634/nmstate-operator/0.log" Jan 26 09:40:06 crc kubenswrapper[4872]: I0126 09:40:06.699526 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-8474b5b9d8-8cgjv_3020f909-4524-4ae9-9fcb-5194b45565de/nmstate-webhook/0.log" Jan 26 09:40:34 crc kubenswrapper[4872]: I0126 09:40:34.778638 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-whrlc_1326511a-69a9-48da-850a-c05a3fa98fad/kube-rbac-proxy/0.log" Jan 26 09:40:34 crc kubenswrapper[4872]: I0126 09:40:34.837035 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6968d8fdc4-whrlc_1326511a-69a9-48da-850a-c05a3fa98fad/controller/0.log" Jan 26 09:40:34 crc kubenswrapper[4872]: I0126 09:40:34.909083 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7df86c4f6c-m7pqv_965facdf-05ce-4a41-8a03-cf20d04932d3/frr-k8s-webhook-server/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.013365 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/cp-frr-files/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.232551 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/cp-frr-files/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.245190 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/cp-reloader/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.260736 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/cp-metrics/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.273976 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/cp-reloader/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.417972 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/cp-frr-files/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.421772 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/cp-reloader/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.435770 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/cp-metrics/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.449121 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/cp-metrics/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.650938 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/cp-reloader/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.663329 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/cp-frr-files/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.670293 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/cp-metrics/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.677333 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/controller/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.856731 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/frr-metrics/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.873033 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/kube-rbac-proxy/0.log" Jan 26 09:40:35 crc kubenswrapper[4872]: I0126 09:40:35.939448 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/kube-rbac-proxy-frr/0.log" Jan 26 09:40:36 crc kubenswrapper[4872]: I0126 09:40:36.150625 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/reloader/0.log" Jan 26 09:40:36 crc kubenswrapper[4872]: I0126 09:40:36.193835 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-56ccf68bf4-prss2_a9dcbe6d-6d0e-4a61-b4c5-9dd0b185496e/manager/0.log" Jan 26 09:40:36 crc kubenswrapper[4872]: I0126 09:40:36.431981 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6d8c8779-66lg8_795e65d7-69c2-4449-b99d-4bbee023b065/webhook-server/0.log" Jan 26 09:40:36 crc kubenswrapper[4872]: I0126 09:40:36.623666 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-c2rkk_816522a7-9120-431e-9f3d-2c66f61d050c/kube-rbac-proxy/0.log" Jan 26 09:40:36 crc kubenswrapper[4872]: I0126 09:40:36.928189 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-zmwfb_bc48ed53-17c1-43f3-9b31-d04c7d510282/frr/0.log" Jan 26 09:40:36 crc kubenswrapper[4872]: I0126 09:40:36.983151 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-c2rkk_816522a7-9120-431e-9f3d-2c66f61d050c/speaker/0.log" Jan 26 09:40:49 crc kubenswrapper[4872]: I0126 09:40:49.497215 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68_18a3643b-548e-4f24-9489-7271c5d5dfee/util/0.log" Jan 26 09:40:49 crc kubenswrapper[4872]: I0126 09:40:49.663833 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68_18a3643b-548e-4f24-9489-7271c5d5dfee/pull/0.log" Jan 26 09:40:49 crc kubenswrapper[4872]: I0126 09:40:49.699879 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68_18a3643b-548e-4f24-9489-7271c5d5dfee/util/0.log" Jan 26 09:40:49 crc kubenswrapper[4872]: I0126 09:40:49.756995 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68_18a3643b-548e-4f24-9489-7271c5d5dfee/pull/0.log" Jan 26 09:40:49 crc kubenswrapper[4872]: I0126 09:40:49.920974 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68_18a3643b-548e-4f24-9489-7271c5d5dfee/pull/0.log" Jan 26 09:40:49 crc kubenswrapper[4872]: I0126 09:40:49.953097 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68_18a3643b-548e-4f24-9489-7271c5d5dfee/extract/0.log" Jan 26 09:40:49 crc kubenswrapper[4872]: I0126 09:40:49.957871 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_270996307cd21d144be796860235064b5127c2fcf62ccccd6689c259dcmfd68_18a3643b-548e-4f24-9489-7271c5d5dfee/util/0.log" Jan 26 09:40:50 crc kubenswrapper[4872]: I0126 09:40:50.106234 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb_077c7b6a-93a7-4851-84fe-549a78533721/util/0.log" Jan 26 09:40:50 crc kubenswrapper[4872]: I0126 09:40:50.268444 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb_077c7b6a-93a7-4851-84fe-549a78533721/util/0.log" Jan 26 09:40:50 crc kubenswrapper[4872]: I0126 09:40:50.278900 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb_077c7b6a-93a7-4851-84fe-549a78533721/pull/0.log" Jan 26 09:40:50 crc kubenswrapper[4872]: I0126 09:40:50.289778 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb_077c7b6a-93a7-4851-84fe-549a78533721/pull/0.log" Jan 26 09:40:50 crc kubenswrapper[4872]: I0126 09:40:50.427731 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb_077c7b6a-93a7-4851-84fe-549a78533721/pull/0.log" Jan 26 09:40:50 crc kubenswrapper[4872]: I0126 09:40:50.439295 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb_077c7b6a-93a7-4851-84fe-549a78533721/extract/0.log" Jan 26 09:40:50 crc kubenswrapper[4872]: I0126 09:40:50.519651 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_53efe8611d43ac2275911d954e05efbbba7920a530aff9253ed1cec7134tzrb_077c7b6a-93a7-4851-84fe-549a78533721/util/0.log" Jan 26 09:40:50 crc kubenswrapper[4872]: I0126 09:40:50.605601 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5mk2s_5f25adda-7523-4fba-9f9e-32583f0d9ad6/extract-utilities/0.log" Jan 26 09:40:50 crc kubenswrapper[4872]: I0126 09:40:50.819274 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5mk2s_5f25adda-7523-4fba-9f9e-32583f0d9ad6/extract-content/0.log" Jan 26 09:40:50 crc kubenswrapper[4872]: I0126 09:40:50.821688 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5mk2s_5f25adda-7523-4fba-9f9e-32583f0d9ad6/extract-content/0.log" Jan 26 09:40:50 crc kubenswrapper[4872]: I0126 09:40:50.832970 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5mk2s_5f25adda-7523-4fba-9f9e-32583f0d9ad6/extract-utilities/0.log" Jan 26 09:40:50 crc kubenswrapper[4872]: I0126 09:40:50.947767 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5mk2s_5f25adda-7523-4fba-9f9e-32583f0d9ad6/extract-utilities/0.log" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.028000 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5mk2s_5f25adda-7523-4fba-9f9e-32583f0d9ad6/extract-content/0.log" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.236349 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xqxmn"] Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.238585 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjj7m_faac185b-bdff-4f6d-801a-1458123a20b0/extract-utilities/0.log" Jan 26 09:40:51 crc kubenswrapper[4872]: E0126 09:40:51.243070 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d126c8a9-3775-448e-9fda-97c60de87d35" containerName="container-00" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.243109 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d126c8a9-3775-448e-9fda-97c60de87d35" containerName="container-00" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.243498 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d126c8a9-3775-448e-9fda-97c60de87d35" containerName="container-00" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.248511 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.249358 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xqxmn"] Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.379600 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-5mk2s_5f25adda-7523-4fba-9f9e-32583f0d9ad6/registry-server/0.log" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.416811 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpbdr\" (UniqueName: \"kubernetes.io/projected/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-kube-api-access-zpbdr\") pod \"community-operators-xqxmn\" (UID: \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\") " pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.416918 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-utilities\") pod \"community-operators-xqxmn\" (UID: \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\") " pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.416969 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-catalog-content\") pod \"community-operators-xqxmn\" (UID: \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\") " pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.434079 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjj7m_faac185b-bdff-4f6d-801a-1458123a20b0/extract-content/0.log" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.471356 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjj7m_faac185b-bdff-4f6d-801a-1458123a20b0/extract-utilities/0.log" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.471370 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjj7m_faac185b-bdff-4f6d-801a-1458123a20b0/extract-content/0.log" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.518900 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-utilities\") pod \"community-operators-xqxmn\" (UID: \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\") " pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.519008 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-catalog-content\") pod \"community-operators-xqxmn\" (UID: \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\") " pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.519114 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpbdr\" (UniqueName: \"kubernetes.io/projected/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-kube-api-access-zpbdr\") pod \"community-operators-xqxmn\" (UID: \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\") " pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.519499 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-catalog-content\") pod \"community-operators-xqxmn\" (UID: \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\") " pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.519768 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-utilities\") pod \"community-operators-xqxmn\" (UID: \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\") " pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.537127 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpbdr\" (UniqueName: \"kubernetes.io/projected/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-kube-api-access-zpbdr\") pod \"community-operators-xqxmn\" (UID: \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\") " pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.588279 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.745749 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjj7m_faac185b-bdff-4f6d-801a-1458123a20b0/extract-content/0.log" Jan 26 09:40:51 crc kubenswrapper[4872]: I0126 09:40:51.830960 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjj7m_faac185b-bdff-4f6d-801a-1458123a20b0/extract-utilities/0.log" Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.076278 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-9ckvb_579725e4-42fb-49f8-b445-4eed08a373b6/marketplace-operator/0.log" Jan 26 09:40:52 crc kubenswrapper[4872]: W0126 09:40:52.147454 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbeba94c2_f00a_45bf_a73f_c7a61bbf92d5.slice/crio-ca6a6d5445b07384263566d019b9d954b05c7edc7e97df154eee6b0fc9e3c4bb WatchSource:0}: Error finding container ca6a6d5445b07384263566d019b9d954b05c7edc7e97df154eee6b0fc9e3c4bb: Status 404 returned error can't find the container with id ca6a6d5445b07384263566d019b9d954b05c7edc7e97df154eee6b0fc9e3c4bb Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.150697 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xqxmn"] Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.166262 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnmnc_41ca443e-2b21-4369-851c-30b312340df2/extract-utilities/0.log" Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.244632 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rjj7m_faac185b-bdff-4f6d-801a-1458123a20b0/registry-server/0.log" Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.324089 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnmnc_41ca443e-2b21-4369-851c-30b312340df2/extract-utilities/0.log" Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.354218 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnmnc_41ca443e-2b21-4369-851c-30b312340df2/extract-content/0.log" Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.362992 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnmnc_41ca443e-2b21-4369-851c-30b312340df2/extract-content/0.log" Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.509889 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnmnc_41ca443e-2b21-4369-851c-30b312340df2/extract-utilities/0.log" Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.547300 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnmnc_41ca443e-2b21-4369-851c-30b312340df2/extract-content/0.log" Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.775071 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-rnmnc_41ca443e-2b21-4369-851c-30b312340df2/registry-server/0.log" Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.822971 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d5qrt_fccd3bc2-18ff-40d4-b862-7257f6cebc5b/extract-utilities/0.log" Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.897164 4872 generic.go:334] "Generic (PLEG): container finished" podID="beba94c2-f00a-45bf-a73f-c7a61bbf92d5" containerID="e2cd58336e1e7f1645e3ad4caf296b1725e2c2da3962a0f4458a5696f57b6b59" exitCode=0 Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.897219 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqxmn" event={"ID":"beba94c2-f00a-45bf-a73f-c7a61bbf92d5","Type":"ContainerDied","Data":"e2cd58336e1e7f1645e3ad4caf296b1725e2c2da3962a0f4458a5696f57b6b59"} Jan 26 09:40:52 crc kubenswrapper[4872]: I0126 09:40:52.897249 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqxmn" event={"ID":"beba94c2-f00a-45bf-a73f-c7a61bbf92d5","Type":"ContainerStarted","Data":"ca6a6d5445b07384263566d019b9d954b05c7edc7e97df154eee6b0fc9e3c4bb"} Jan 26 09:40:53 crc kubenswrapper[4872]: I0126 09:40:53.006631 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d5qrt_fccd3bc2-18ff-40d4-b862-7257f6cebc5b/extract-content/0.log" Jan 26 09:40:53 crc kubenswrapper[4872]: I0126 09:40:53.007015 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d5qrt_fccd3bc2-18ff-40d4-b862-7257f6cebc5b/extract-content/0.log" Jan 26 09:40:53 crc kubenswrapper[4872]: I0126 09:40:53.021891 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d5qrt_fccd3bc2-18ff-40d4-b862-7257f6cebc5b/extract-utilities/0.log" Jan 26 09:40:53 crc kubenswrapper[4872]: I0126 09:40:53.191240 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d5qrt_fccd3bc2-18ff-40d4-b862-7257f6cebc5b/extract-utilities/0.log" Jan 26 09:40:53 crc kubenswrapper[4872]: I0126 09:40:53.238647 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d5qrt_fccd3bc2-18ff-40d4-b862-7257f6cebc5b/extract-content/0.log" Jan 26 09:40:53 crc kubenswrapper[4872]: I0126 09:40:53.530137 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-d5qrt_fccd3bc2-18ff-40d4-b862-7257f6cebc5b/registry-server/0.log" Jan 26 09:40:54 crc kubenswrapper[4872]: I0126 09:40:54.918789 4872 generic.go:334] "Generic (PLEG): container finished" podID="beba94c2-f00a-45bf-a73f-c7a61bbf92d5" containerID="700b6881d1e2bae49ed11d341a6a18c9c5365a3bcb418ca3737e91a6d954422c" exitCode=0 Jan 26 09:40:54 crc kubenswrapper[4872]: I0126 09:40:54.918882 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqxmn" event={"ID":"beba94c2-f00a-45bf-a73f-c7a61bbf92d5","Type":"ContainerDied","Data":"700b6881d1e2bae49ed11d341a6a18c9c5365a3bcb418ca3737e91a6d954422c"} Jan 26 09:40:55 crc kubenswrapper[4872]: I0126 09:40:55.928182 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqxmn" event={"ID":"beba94c2-f00a-45bf-a73f-c7a61bbf92d5","Type":"ContainerStarted","Data":"3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56"} Jan 26 09:40:55 crc kubenswrapper[4872]: I0126 09:40:55.954162 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xqxmn" podStartSLOduration=2.295368768 podStartE2EDuration="4.954143254s" podCreationTimestamp="2026-01-26 09:40:51 +0000 UTC" firstStartedPulling="2026-01-26 09:40:52.899395506 +0000 UTC m=+1986.208235307" lastFinishedPulling="2026-01-26 09:40:55.558169992 +0000 UTC m=+1988.867009793" observedRunningTime="2026-01-26 09:40:55.947480502 +0000 UTC m=+1989.256320303" watchObservedRunningTime="2026-01-26 09:40:55.954143254 +0000 UTC m=+1989.262983055" Jan 26 09:41:01 crc kubenswrapper[4872]: I0126 09:41:01.588653 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:41:01 crc kubenswrapper[4872]: I0126 09:41:01.589710 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:41:01 crc kubenswrapper[4872]: I0126 09:41:01.645560 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:41:02 crc kubenswrapper[4872]: I0126 09:41:02.052237 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:41:02 crc kubenswrapper[4872]: I0126 09:41:02.116748 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xqxmn"] Jan 26 09:41:04 crc kubenswrapper[4872]: I0126 09:41:04.013510 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xqxmn" podUID="beba94c2-f00a-45bf-a73f-c7a61bbf92d5" containerName="registry-server" containerID="cri-o://3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56" gracePeriod=2 Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.009997 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.025561 4872 generic.go:334] "Generic (PLEG): container finished" podID="beba94c2-f00a-45bf-a73f-c7a61bbf92d5" containerID="3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56" exitCode=0 Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.025604 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqxmn" event={"ID":"beba94c2-f00a-45bf-a73f-c7a61bbf92d5","Type":"ContainerDied","Data":"3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56"} Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.025630 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xqxmn" event={"ID":"beba94c2-f00a-45bf-a73f-c7a61bbf92d5","Type":"ContainerDied","Data":"ca6a6d5445b07384263566d019b9d954b05c7edc7e97df154eee6b0fc9e3c4bb"} Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.025642 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xqxmn" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.025648 4872 scope.go:117] "RemoveContainer" containerID="3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.066130 4872 scope.go:117] "RemoveContainer" containerID="700b6881d1e2bae49ed11d341a6a18c9c5365a3bcb418ca3737e91a6d954422c" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.094602 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpbdr\" (UniqueName: \"kubernetes.io/projected/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-kube-api-access-zpbdr\") pod \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\" (UID: \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\") " Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.094647 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-catalog-content\") pod \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\" (UID: \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\") " Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.094772 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-utilities\") pod \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\" (UID: \"beba94c2-f00a-45bf-a73f-c7a61bbf92d5\") " Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.098533 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-utilities" (OuterVolumeSpecName: "utilities") pod "beba94c2-f00a-45bf-a73f-c7a61bbf92d5" (UID: "beba94c2-f00a-45bf-a73f-c7a61bbf92d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.102014 4872 scope.go:117] "RemoveContainer" containerID="e2cd58336e1e7f1645e3ad4caf296b1725e2c2da3962a0f4458a5696f57b6b59" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.106069 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-kube-api-access-zpbdr" (OuterVolumeSpecName: "kube-api-access-zpbdr") pod "beba94c2-f00a-45bf-a73f-c7a61bbf92d5" (UID: "beba94c2-f00a-45bf-a73f-c7a61bbf92d5"). InnerVolumeSpecName "kube-api-access-zpbdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.169486 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "beba94c2-f00a-45bf-a73f-c7a61bbf92d5" (UID: "beba94c2-f00a-45bf-a73f-c7a61bbf92d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.179053 4872 scope.go:117] "RemoveContainer" containerID="3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56" Jan 26 09:41:05 crc kubenswrapper[4872]: E0126 09:41:05.181295 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56\": container with ID starting with 3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56 not found: ID does not exist" containerID="3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.181337 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56"} err="failed to get container status \"3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56\": rpc error: code = NotFound desc = could not find container \"3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56\": container with ID starting with 3615a5148f0e561d036bfc7ff06591afc58a83b358a68d0059bfeed4c4a05e56 not found: ID does not exist" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.181369 4872 scope.go:117] "RemoveContainer" containerID="700b6881d1e2bae49ed11d341a6a18c9c5365a3bcb418ca3737e91a6d954422c" Jan 26 09:41:05 crc kubenswrapper[4872]: E0126 09:41:05.182013 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"700b6881d1e2bae49ed11d341a6a18c9c5365a3bcb418ca3737e91a6d954422c\": container with ID starting with 700b6881d1e2bae49ed11d341a6a18c9c5365a3bcb418ca3737e91a6d954422c not found: ID does not exist" containerID="700b6881d1e2bae49ed11d341a6a18c9c5365a3bcb418ca3737e91a6d954422c" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.182043 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"700b6881d1e2bae49ed11d341a6a18c9c5365a3bcb418ca3737e91a6d954422c"} err="failed to get container status \"700b6881d1e2bae49ed11d341a6a18c9c5365a3bcb418ca3737e91a6d954422c\": rpc error: code = NotFound desc = could not find container \"700b6881d1e2bae49ed11d341a6a18c9c5365a3bcb418ca3737e91a6d954422c\": container with ID starting with 700b6881d1e2bae49ed11d341a6a18c9c5365a3bcb418ca3737e91a6d954422c not found: ID does not exist" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.182069 4872 scope.go:117] "RemoveContainer" containerID="e2cd58336e1e7f1645e3ad4caf296b1725e2c2da3962a0f4458a5696f57b6b59" Jan 26 09:41:05 crc kubenswrapper[4872]: E0126 09:41:05.182353 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2cd58336e1e7f1645e3ad4caf296b1725e2c2da3962a0f4458a5696f57b6b59\": container with ID starting with e2cd58336e1e7f1645e3ad4caf296b1725e2c2da3962a0f4458a5696f57b6b59 not found: ID does not exist" containerID="e2cd58336e1e7f1645e3ad4caf296b1725e2c2da3962a0f4458a5696f57b6b59" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.182390 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2cd58336e1e7f1645e3ad4caf296b1725e2c2da3962a0f4458a5696f57b6b59"} err="failed to get container status \"e2cd58336e1e7f1645e3ad4caf296b1725e2c2da3962a0f4458a5696f57b6b59\": rpc error: code = NotFound desc = could not find container \"e2cd58336e1e7f1645e3ad4caf296b1725e2c2da3962a0f4458a5696f57b6b59\": container with ID starting with e2cd58336e1e7f1645e3ad4caf296b1725e2c2da3962a0f4458a5696f57b6b59 not found: ID does not exist" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.197102 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpbdr\" (UniqueName: \"kubernetes.io/projected/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-kube-api-access-zpbdr\") on node \"crc\" DevicePath \"\"" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.197136 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.197148 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/beba94c2-f00a-45bf-a73f-c7a61bbf92d5-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.345256 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xqxmn"] Jan 26 09:41:05 crc kubenswrapper[4872]: I0126 09:41:05.351164 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xqxmn"] Jan 26 09:41:07 crc kubenswrapper[4872]: I0126 09:41:07.201773 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beba94c2-f00a-45bf-a73f-c7a61bbf92d5" path="/var/lib/kubelet/pods/beba94c2-f00a-45bf-a73f-c7a61bbf92d5/volumes" Jan 26 09:42:10 crc kubenswrapper[4872]: I0126 09:42:10.242201 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:42:10 crc kubenswrapper[4872]: I0126 09:42:10.242883 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:42:28 crc kubenswrapper[4872]: I0126 09:42:28.920685 4872 generic.go:334] "Generic (PLEG): container finished" podID="273d271b-605b-4dd7-b4d1-4046f63cef78" containerID="c08c21452faf92710d7919aae65fdb95a33085480e9d7020c8deac1be4b76925" exitCode=0 Jan 26 09:42:28 crc kubenswrapper[4872]: I0126 09:42:28.920760 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-6kdtd/must-gather-wfp8g" event={"ID":"273d271b-605b-4dd7-b4d1-4046f63cef78","Type":"ContainerDied","Data":"c08c21452faf92710d7919aae65fdb95a33085480e9d7020c8deac1be4b76925"} Jan 26 09:42:28 crc kubenswrapper[4872]: I0126 09:42:28.921788 4872 scope.go:117] "RemoveContainer" containerID="c08c21452faf92710d7919aae65fdb95a33085480e9d7020c8deac1be4b76925" Jan 26 09:42:29 crc kubenswrapper[4872]: I0126 09:42:29.068633 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6kdtd_must-gather-wfp8g_273d271b-605b-4dd7-b4d1-4046f63cef78/gather/0.log" Jan 26 09:42:37 crc kubenswrapper[4872]: I0126 09:42:37.160711 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-6kdtd/must-gather-wfp8g"] Jan 26 09:42:37 crc kubenswrapper[4872]: I0126 09:42:37.161577 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-6kdtd/must-gather-wfp8g" podUID="273d271b-605b-4dd7-b4d1-4046f63cef78" containerName="copy" containerID="cri-o://fb717a9caf98e4c05515f00cdc6e43d955a9a62f1062688a548d89f2dcaeb5d9" gracePeriod=2 Jan 26 09:42:37 crc kubenswrapper[4872]: I0126 09:42:37.168795 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-6kdtd/must-gather-wfp8g"] Jan 26 09:42:38 crc kubenswrapper[4872]: I0126 09:42:38.006117 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6kdtd_must-gather-wfp8g_273d271b-605b-4dd7-b4d1-4046f63cef78/copy/0.log" Jan 26 09:42:38 crc kubenswrapper[4872]: I0126 09:42:38.006638 4872 generic.go:334] "Generic (PLEG): container finished" podID="273d271b-605b-4dd7-b4d1-4046f63cef78" containerID="fb717a9caf98e4c05515f00cdc6e43d955a9a62f1062688a548d89f2dcaeb5d9" exitCode=143 Jan 26 09:42:38 crc kubenswrapper[4872]: I0126 09:42:38.133975 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6kdtd_must-gather-wfp8g_273d271b-605b-4dd7-b4d1-4046f63cef78/copy/0.log" Jan 26 09:42:38 crc kubenswrapper[4872]: I0126 09:42:38.135198 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/must-gather-wfp8g" Jan 26 09:42:38 crc kubenswrapper[4872]: I0126 09:42:38.217093 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z47tk\" (UniqueName: \"kubernetes.io/projected/273d271b-605b-4dd7-b4d1-4046f63cef78-kube-api-access-z47tk\") pod \"273d271b-605b-4dd7-b4d1-4046f63cef78\" (UID: \"273d271b-605b-4dd7-b4d1-4046f63cef78\") " Jan 26 09:42:38 crc kubenswrapper[4872]: I0126 09:42:38.217176 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/273d271b-605b-4dd7-b4d1-4046f63cef78-must-gather-output\") pod \"273d271b-605b-4dd7-b4d1-4046f63cef78\" (UID: \"273d271b-605b-4dd7-b4d1-4046f63cef78\") " Jan 26 09:42:38 crc kubenswrapper[4872]: I0126 09:42:38.225475 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/273d271b-605b-4dd7-b4d1-4046f63cef78-kube-api-access-z47tk" (OuterVolumeSpecName: "kube-api-access-z47tk") pod "273d271b-605b-4dd7-b4d1-4046f63cef78" (UID: "273d271b-605b-4dd7-b4d1-4046f63cef78"). InnerVolumeSpecName "kube-api-access-z47tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:42:38 crc kubenswrapper[4872]: I0126 09:42:38.319369 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z47tk\" (UniqueName: \"kubernetes.io/projected/273d271b-605b-4dd7-b4d1-4046f63cef78-kube-api-access-z47tk\") on node \"crc\" DevicePath \"\"" Jan 26 09:42:38 crc kubenswrapper[4872]: I0126 09:42:38.384529 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/273d271b-605b-4dd7-b4d1-4046f63cef78-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "273d271b-605b-4dd7-b4d1-4046f63cef78" (UID: "273d271b-605b-4dd7-b4d1-4046f63cef78"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:42:38 crc kubenswrapper[4872]: I0126 09:42:38.420886 4872 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/273d271b-605b-4dd7-b4d1-4046f63cef78-must-gather-output\") on node \"crc\" DevicePath \"\"" Jan 26 09:42:39 crc kubenswrapper[4872]: I0126 09:42:39.018050 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-6kdtd_must-gather-wfp8g_273d271b-605b-4dd7-b4d1-4046f63cef78/copy/0.log" Jan 26 09:42:39 crc kubenswrapper[4872]: I0126 09:42:39.018567 4872 scope.go:117] "RemoveContainer" containerID="fb717a9caf98e4c05515f00cdc6e43d955a9a62f1062688a548d89f2dcaeb5d9" Jan 26 09:42:39 crc kubenswrapper[4872]: I0126 09:42:39.018680 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-6kdtd/must-gather-wfp8g" Jan 26 09:42:39 crc kubenswrapper[4872]: I0126 09:42:39.046516 4872 scope.go:117] "RemoveContainer" containerID="c08c21452faf92710d7919aae65fdb95a33085480e9d7020c8deac1be4b76925" Jan 26 09:42:39 crc kubenswrapper[4872]: I0126 09:42:39.194718 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="273d271b-605b-4dd7-b4d1-4046f63cef78" path="/var/lib/kubelet/pods/273d271b-605b-4dd7-b4d1-4046f63cef78/volumes" Jan 26 09:42:40 crc kubenswrapper[4872]: I0126 09:42:40.227968 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:42:40 crc kubenswrapper[4872]: I0126 09:42:40.228062 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.259899 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lnmtf"] Jan 26 09:42:42 crc kubenswrapper[4872]: E0126 09:42:42.260495 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="273d271b-605b-4dd7-b4d1-4046f63cef78" containerName="copy" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.260518 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="273d271b-605b-4dd7-b4d1-4046f63cef78" containerName="copy" Jan 26 09:42:42 crc kubenswrapper[4872]: E0126 09:42:42.260559 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beba94c2-f00a-45bf-a73f-c7a61bbf92d5" containerName="extract-content" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.260573 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="beba94c2-f00a-45bf-a73f-c7a61bbf92d5" containerName="extract-content" Jan 26 09:42:42 crc kubenswrapper[4872]: E0126 09:42:42.260599 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beba94c2-f00a-45bf-a73f-c7a61bbf92d5" containerName="registry-server" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.260611 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="beba94c2-f00a-45bf-a73f-c7a61bbf92d5" containerName="registry-server" Jan 26 09:42:42 crc kubenswrapper[4872]: E0126 09:42:42.260643 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beba94c2-f00a-45bf-a73f-c7a61bbf92d5" containerName="extract-utilities" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.260655 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="beba94c2-f00a-45bf-a73f-c7a61bbf92d5" containerName="extract-utilities" Jan 26 09:42:42 crc kubenswrapper[4872]: E0126 09:42:42.260682 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="273d271b-605b-4dd7-b4d1-4046f63cef78" containerName="gather" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.260694 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="273d271b-605b-4dd7-b4d1-4046f63cef78" containerName="gather" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.261057 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="beba94c2-f00a-45bf-a73f-c7a61bbf92d5" containerName="registry-server" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.261085 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="273d271b-605b-4dd7-b4d1-4046f63cef78" containerName="copy" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.261114 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="273d271b-605b-4dd7-b4d1-4046f63cef78" containerName="gather" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.263508 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.273683 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnmtf"] Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.400014 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gwxp\" (UniqueName: \"kubernetes.io/projected/a5e2fbd8-de94-467d-9d73-2139b58f43a5-kube-api-access-6gwxp\") pod \"redhat-marketplace-lnmtf\" (UID: \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\") " pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.400097 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e2fbd8-de94-467d-9d73-2139b58f43a5-utilities\") pod \"redhat-marketplace-lnmtf\" (UID: \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\") " pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.400213 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e2fbd8-de94-467d-9d73-2139b58f43a5-catalog-content\") pod \"redhat-marketplace-lnmtf\" (UID: \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\") " pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.501635 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gwxp\" (UniqueName: \"kubernetes.io/projected/a5e2fbd8-de94-467d-9d73-2139b58f43a5-kube-api-access-6gwxp\") pod \"redhat-marketplace-lnmtf\" (UID: \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\") " pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.501707 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e2fbd8-de94-467d-9d73-2139b58f43a5-utilities\") pod \"redhat-marketplace-lnmtf\" (UID: \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\") " pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.501743 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e2fbd8-de94-467d-9d73-2139b58f43a5-catalog-content\") pod \"redhat-marketplace-lnmtf\" (UID: \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\") " pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.502270 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e2fbd8-de94-467d-9d73-2139b58f43a5-catalog-content\") pod \"redhat-marketplace-lnmtf\" (UID: \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\") " pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.502461 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e2fbd8-de94-467d-9d73-2139b58f43a5-utilities\") pod \"redhat-marketplace-lnmtf\" (UID: \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\") " pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.523429 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gwxp\" (UniqueName: \"kubernetes.io/projected/a5e2fbd8-de94-467d-9d73-2139b58f43a5-kube-api-access-6gwxp\") pod \"redhat-marketplace-lnmtf\" (UID: \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\") " pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:42 crc kubenswrapper[4872]: I0126 09:42:42.589339 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:43 crc kubenswrapper[4872]: I0126 09:42:43.039765 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnmtf"] Jan 26 09:42:43 crc kubenswrapper[4872]: I0126 09:42:43.056632 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnmtf" event={"ID":"a5e2fbd8-de94-467d-9d73-2139b58f43a5","Type":"ContainerStarted","Data":"c40a37ec7eb318841244fc75e143aef5cdfd44b1a3f53f733ee8f799e393877e"} Jan 26 09:42:43 crc kubenswrapper[4872]: E0126 09:42:43.363687 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5e2fbd8_de94_467d_9d73_2139b58f43a5.slice/crio-conmon-6ce3a674b4b74ac7837f41ec120706f30c33c80c397b15790575492a41a5595f.scope\": RecentStats: unable to find data in memory cache]" Jan 26 09:42:44 crc kubenswrapper[4872]: I0126 09:42:44.073269 4872 generic.go:334] "Generic (PLEG): container finished" podID="a5e2fbd8-de94-467d-9d73-2139b58f43a5" containerID="6ce3a674b4b74ac7837f41ec120706f30c33c80c397b15790575492a41a5595f" exitCode=0 Jan 26 09:42:44 crc kubenswrapper[4872]: I0126 09:42:44.073633 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnmtf" event={"ID":"a5e2fbd8-de94-467d-9d73-2139b58f43a5","Type":"ContainerDied","Data":"6ce3a674b4b74ac7837f41ec120706f30c33c80c397b15790575492a41a5595f"} Jan 26 09:42:44 crc kubenswrapper[4872]: I0126 09:42:44.079197 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Jan 26 09:42:45 crc kubenswrapper[4872]: I0126 09:42:45.127777 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnmtf" event={"ID":"a5e2fbd8-de94-467d-9d73-2139b58f43a5","Type":"ContainerStarted","Data":"24d635d1c1560ec54e3d4299a86075764fffb4be883d0df79b569d3b6ba300c3"} Jan 26 09:42:46 crc kubenswrapper[4872]: I0126 09:42:46.171166 4872 generic.go:334] "Generic (PLEG): container finished" podID="a5e2fbd8-de94-467d-9d73-2139b58f43a5" containerID="24d635d1c1560ec54e3d4299a86075764fffb4be883d0df79b569d3b6ba300c3" exitCode=0 Jan 26 09:42:46 crc kubenswrapper[4872]: I0126 09:42:46.171607 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnmtf" event={"ID":"a5e2fbd8-de94-467d-9d73-2139b58f43a5","Type":"ContainerDied","Data":"24d635d1c1560ec54e3d4299a86075764fffb4be883d0df79b569d3b6ba300c3"} Jan 26 09:42:47 crc kubenswrapper[4872]: I0126 09:42:47.201336 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnmtf" event={"ID":"a5e2fbd8-de94-467d-9d73-2139b58f43a5","Type":"ContainerStarted","Data":"71e5ba30c636d7512c79f5f1e698da7e9e68a0a5b0ae1bd1ecd38bfc83e9f624"} Jan 26 09:42:47 crc kubenswrapper[4872]: I0126 09:42:47.214174 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lnmtf" podStartSLOduration=2.694948174 podStartE2EDuration="5.214150948s" podCreationTimestamp="2026-01-26 09:42:42 +0000 UTC" firstStartedPulling="2026-01-26 09:42:44.078471495 +0000 UTC m=+2097.387311336" lastFinishedPulling="2026-01-26 09:42:46.597674289 +0000 UTC m=+2099.906514110" observedRunningTime="2026-01-26 09:42:47.210536191 +0000 UTC m=+2100.519375992" watchObservedRunningTime="2026-01-26 09:42:47.214150948 +0000 UTC m=+2100.522990749" Jan 26 09:42:52 crc kubenswrapper[4872]: I0126 09:42:52.589775 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:52 crc kubenswrapper[4872]: I0126 09:42:52.592694 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:52 crc kubenswrapper[4872]: I0126 09:42:52.652721 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:53 crc kubenswrapper[4872]: I0126 09:42:53.309135 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:53 crc kubenswrapper[4872]: I0126 09:42:53.377363 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnmtf"] Jan 26 09:42:55 crc kubenswrapper[4872]: I0126 09:42:55.259475 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lnmtf" podUID="a5e2fbd8-de94-467d-9d73-2139b58f43a5" containerName="registry-server" containerID="cri-o://71e5ba30c636d7512c79f5f1e698da7e9e68a0a5b0ae1bd1ecd38bfc83e9f624" gracePeriod=2 Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.270588 4872 generic.go:334] "Generic (PLEG): container finished" podID="a5e2fbd8-de94-467d-9d73-2139b58f43a5" containerID="71e5ba30c636d7512c79f5f1e698da7e9e68a0a5b0ae1bd1ecd38bfc83e9f624" exitCode=0 Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.271068 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnmtf" event={"ID":"a5e2fbd8-de94-467d-9d73-2139b58f43a5","Type":"ContainerDied","Data":"71e5ba30c636d7512c79f5f1e698da7e9e68a0a5b0ae1bd1ecd38bfc83e9f624"} Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.271156 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnmtf" event={"ID":"a5e2fbd8-de94-467d-9d73-2139b58f43a5","Type":"ContainerDied","Data":"c40a37ec7eb318841244fc75e143aef5cdfd44b1a3f53f733ee8f799e393877e"} Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.271173 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c40a37ec7eb318841244fc75e143aef5cdfd44b1a3f53f733ee8f799e393877e" Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.322389 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.399313 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gwxp\" (UniqueName: \"kubernetes.io/projected/a5e2fbd8-de94-467d-9d73-2139b58f43a5-kube-api-access-6gwxp\") pod \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\" (UID: \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\") " Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.399476 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e2fbd8-de94-467d-9d73-2139b58f43a5-utilities\") pod \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\" (UID: \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\") " Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.399646 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e2fbd8-de94-467d-9d73-2139b58f43a5-catalog-content\") pod \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\" (UID: \"a5e2fbd8-de94-467d-9d73-2139b58f43a5\") " Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.400841 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e2fbd8-de94-467d-9d73-2139b58f43a5-utilities" (OuterVolumeSpecName: "utilities") pod "a5e2fbd8-de94-467d-9d73-2139b58f43a5" (UID: "a5e2fbd8-de94-467d-9d73-2139b58f43a5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.405309 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5e2fbd8-de94-467d-9d73-2139b58f43a5-kube-api-access-6gwxp" (OuterVolumeSpecName: "kube-api-access-6gwxp") pod "a5e2fbd8-de94-467d-9d73-2139b58f43a5" (UID: "a5e2fbd8-de94-467d-9d73-2139b58f43a5"). InnerVolumeSpecName "kube-api-access-6gwxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.421374 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5e2fbd8-de94-467d-9d73-2139b58f43a5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5e2fbd8-de94-467d-9d73-2139b58f43a5" (UID: "a5e2fbd8-de94-467d-9d73-2139b58f43a5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.501586 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gwxp\" (UniqueName: \"kubernetes.io/projected/a5e2fbd8-de94-467d-9d73-2139b58f43a5-kube-api-access-6gwxp\") on node \"crc\" DevicePath \"\"" Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.501650 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5e2fbd8-de94-467d-9d73-2139b58f43a5-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:42:56 crc kubenswrapper[4872]: I0126 09:42:56.501660 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5e2fbd8-de94-467d-9d73-2139b58f43a5-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:42:57 crc kubenswrapper[4872]: I0126 09:42:57.281019 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnmtf" Jan 26 09:42:57 crc kubenswrapper[4872]: I0126 09:42:57.324652 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnmtf"] Jan 26 09:42:57 crc kubenswrapper[4872]: I0126 09:42:57.331857 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnmtf"] Jan 26 09:42:59 crc kubenswrapper[4872]: I0126 09:42:59.194679 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5e2fbd8-de94-467d-9d73-2139b58f43a5" path="/var/lib/kubelet/pods/a5e2fbd8-de94-467d-9d73-2139b58f43a5/volumes" Jan 26 09:43:10 crc kubenswrapper[4872]: I0126 09:43:10.228319 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:43:10 crc kubenswrapper[4872]: I0126 09:43:10.228983 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:43:10 crc kubenswrapper[4872]: I0126 09:43:10.229037 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:43:10 crc kubenswrapper[4872]: I0126 09:43:10.229879 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8dad67da6bdcbaf03f302de0fdcef76cd940f73a0eab7a62b9aac33c1f3c41aa"} pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 09:43:10 crc kubenswrapper[4872]: I0126 09:43:10.229955 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" containerID="cri-o://8dad67da6bdcbaf03f302de0fdcef76cd940f73a0eab7a62b9aac33c1f3c41aa" gracePeriod=600 Jan 26 09:43:11 crc kubenswrapper[4872]: I0126 09:43:11.411611 4872 generic.go:334] "Generic (PLEG): container finished" podID="fca44d96-a000-4bf2-8283-a937b0192880" containerID="8dad67da6bdcbaf03f302de0fdcef76cd940f73a0eab7a62b9aac33c1f3c41aa" exitCode=0 Jan 26 09:43:11 crc kubenswrapper[4872]: I0126 09:43:11.411677 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerDied","Data":"8dad67da6bdcbaf03f302de0fdcef76cd940f73a0eab7a62b9aac33c1f3c41aa"} Jan 26 09:43:11 crc kubenswrapper[4872]: I0126 09:43:11.412010 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerStarted","Data":"56255d0196d693b2765251d758eac96be427715e4d4c8fd546a448498cf230c2"} Jan 26 09:43:11 crc kubenswrapper[4872]: I0126 09:43:11.412033 4872 scope.go:117] "RemoveContainer" containerID="fab8014eb7f3abb4d2b42b4ee45b559059d669d02102afc52e58c8be683bf32a" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.444669 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2mbq7"] Jan 26 09:43:17 crc kubenswrapper[4872]: E0126 09:43:17.446695 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e2fbd8-de94-467d-9d73-2139b58f43a5" containerName="extract-content" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.446818 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e2fbd8-de94-467d-9d73-2139b58f43a5" containerName="extract-content" Jan 26 09:43:17 crc kubenswrapper[4872]: E0126 09:43:17.446927 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e2fbd8-de94-467d-9d73-2139b58f43a5" containerName="extract-utilities" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.447017 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e2fbd8-de94-467d-9d73-2139b58f43a5" containerName="extract-utilities" Jan 26 09:43:17 crc kubenswrapper[4872]: E0126 09:43:17.447142 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5e2fbd8-de94-467d-9d73-2139b58f43a5" containerName="registry-server" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.447226 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5e2fbd8-de94-467d-9d73-2139b58f43a5" containerName="registry-server" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.447539 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5e2fbd8-de94-467d-9d73-2139b58f43a5" containerName="registry-server" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.449795 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.485991 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2mbq7"] Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.552615 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-utilities\") pod \"certified-operators-2mbq7\" (UID: \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\") " pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.552743 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v49g6\" (UniqueName: \"kubernetes.io/projected/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-kube-api-access-v49g6\") pod \"certified-operators-2mbq7\" (UID: \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\") " pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.552784 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-catalog-content\") pod \"certified-operators-2mbq7\" (UID: \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\") " pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.654856 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v49g6\" (UniqueName: \"kubernetes.io/projected/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-kube-api-access-v49g6\") pod \"certified-operators-2mbq7\" (UID: \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\") " pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.654916 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-catalog-content\") pod \"certified-operators-2mbq7\" (UID: \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\") " pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.655065 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-utilities\") pod \"certified-operators-2mbq7\" (UID: \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\") " pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.656019 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-utilities\") pod \"certified-operators-2mbq7\" (UID: \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\") " pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.656293 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-catalog-content\") pod \"certified-operators-2mbq7\" (UID: \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\") " pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.692554 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v49g6\" (UniqueName: \"kubernetes.io/projected/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-kube-api-access-v49g6\") pod \"certified-operators-2mbq7\" (UID: \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\") " pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:17 crc kubenswrapper[4872]: I0126 09:43:17.773040 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:18 crc kubenswrapper[4872]: I0126 09:43:18.296937 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2mbq7"] Jan 26 09:43:18 crc kubenswrapper[4872]: I0126 09:43:18.508529 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mbq7" event={"ID":"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2","Type":"ContainerStarted","Data":"5dbfcf050d2838693f6d24a834df35ab6738dca4d091e097a0f2bbfc09fdd559"} Jan 26 09:43:18 crc kubenswrapper[4872]: I0126 09:43:18.508603 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mbq7" event={"ID":"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2","Type":"ContainerStarted","Data":"82dca6531b8fc7e4040c7b432833fafd43af8db97a7b902ab993451d46a412f2"} Jan 26 09:43:19 crc kubenswrapper[4872]: I0126 09:43:19.521972 4872 generic.go:334] "Generic (PLEG): container finished" podID="8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" containerID="5dbfcf050d2838693f6d24a834df35ab6738dca4d091e097a0f2bbfc09fdd559" exitCode=0 Jan 26 09:43:19 crc kubenswrapper[4872]: I0126 09:43:19.522076 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mbq7" event={"ID":"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2","Type":"ContainerDied","Data":"5dbfcf050d2838693f6d24a834df35ab6738dca4d091e097a0f2bbfc09fdd559"} Jan 26 09:43:20 crc kubenswrapper[4872]: I0126 09:43:20.532109 4872 generic.go:334] "Generic (PLEG): container finished" podID="8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" containerID="0b915051c8da2b2971a766f79f3ddaf5cc5ecff5e2e69c42e36f11322a4837f4" exitCode=0 Jan 26 09:43:20 crc kubenswrapper[4872]: I0126 09:43:20.532206 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mbq7" event={"ID":"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2","Type":"ContainerDied","Data":"0b915051c8da2b2971a766f79f3ddaf5cc5ecff5e2e69c42e36f11322a4837f4"} Jan 26 09:43:21 crc kubenswrapper[4872]: I0126 09:43:21.545637 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mbq7" event={"ID":"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2","Type":"ContainerStarted","Data":"77fcf3f975c610626314c130d2bfaa0388c349d7cb84afda276ba255e230ab0d"} Jan 26 09:43:21 crc kubenswrapper[4872]: I0126 09:43:21.573045 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2mbq7" podStartSLOduration=3.172294525 podStartE2EDuration="4.573019311s" podCreationTimestamp="2026-01-26 09:43:17 +0000 UTC" firstStartedPulling="2026-01-26 09:43:19.525704624 +0000 UTC m=+2132.834544455" lastFinishedPulling="2026-01-26 09:43:20.92642944 +0000 UTC m=+2134.235269241" observedRunningTime="2026-01-26 09:43:21.56552852 +0000 UTC m=+2134.874368321" watchObservedRunningTime="2026-01-26 09:43:21.573019311 +0000 UTC m=+2134.881859122" Jan 26 09:43:27 crc kubenswrapper[4872]: I0126 09:43:27.773793 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:27 crc kubenswrapper[4872]: I0126 09:43:27.774470 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:27 crc kubenswrapper[4872]: I0126 09:43:27.897559 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:28 crc kubenswrapper[4872]: I0126 09:43:28.701016 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:28 crc kubenswrapper[4872]: I0126 09:43:28.759356 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2mbq7"] Jan 26 09:43:30 crc kubenswrapper[4872]: I0126 09:43:30.666040 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2mbq7" podUID="8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" containerName="registry-server" containerID="cri-o://77fcf3f975c610626314c130d2bfaa0388c349d7cb84afda276ba255e230ab0d" gracePeriod=2 Jan 26 09:43:31 crc kubenswrapper[4872]: I0126 09:43:31.675512 4872 generic.go:334] "Generic (PLEG): container finished" podID="8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" containerID="77fcf3f975c610626314c130d2bfaa0388c349d7cb84afda276ba255e230ab0d" exitCode=0 Jan 26 09:43:31 crc kubenswrapper[4872]: I0126 09:43:31.675617 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mbq7" event={"ID":"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2","Type":"ContainerDied","Data":"77fcf3f975c610626314c130d2bfaa0388c349d7cb84afda276ba255e230ab0d"} Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.364101 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.562969 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v49g6\" (UniqueName: \"kubernetes.io/projected/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-kube-api-access-v49g6\") pod \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\" (UID: \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\") " Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.563138 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-utilities\") pod \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\" (UID: \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\") " Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.563171 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-catalog-content\") pod \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\" (UID: \"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2\") " Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.563977 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-utilities" (OuterVolumeSpecName: "utilities") pod "8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" (UID: "8514cdec-1aa1-4da5-81b9-d4a58dda8bb2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.568687 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-kube-api-access-v49g6" (OuterVolumeSpecName: "kube-api-access-v49g6") pod "8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" (UID: "8514cdec-1aa1-4da5-81b9-d4a58dda8bb2"). InnerVolumeSpecName "kube-api-access-v49g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.605259 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" (UID: "8514cdec-1aa1-4da5-81b9-d4a58dda8bb2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.666021 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-utilities\") on node \"crc\" DevicePath \"\"" Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.666062 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-catalog-content\") on node \"crc\" DevicePath \"\"" Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.666077 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v49g6\" (UniqueName: \"kubernetes.io/projected/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2-kube-api-access-v49g6\") on node \"crc\" DevicePath \"\"" Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.685069 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2mbq7" event={"ID":"8514cdec-1aa1-4da5-81b9-d4a58dda8bb2","Type":"ContainerDied","Data":"82dca6531b8fc7e4040c7b432833fafd43af8db97a7b902ab993451d46a412f2"} Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.685122 4872 scope.go:117] "RemoveContainer" containerID="77fcf3f975c610626314c130d2bfaa0388c349d7cb84afda276ba255e230ab0d" Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.685264 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2mbq7" Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.704053 4872 scope.go:117] "RemoveContainer" containerID="0b915051c8da2b2971a766f79f3ddaf5cc5ecff5e2e69c42e36f11322a4837f4" Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.721154 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2mbq7"] Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.728945 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2mbq7"] Jan 26 09:43:32 crc kubenswrapper[4872]: I0126 09:43:32.739299 4872 scope.go:117] "RemoveContainer" containerID="5dbfcf050d2838693f6d24a834df35ab6738dca4d091e097a0f2bbfc09fdd559" Jan 26 09:43:33 crc kubenswrapper[4872]: I0126 09:43:33.210983 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" path="/var/lib/kubelet/pods/8514cdec-1aa1-4da5-81b9-d4a58dda8bb2/volumes" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.176985 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh"] Jan 26 09:45:00 crc kubenswrapper[4872]: E0126 09:45:00.178292 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" containerName="extract-content" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.178318 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" containerName="extract-content" Jan 26 09:45:00 crc kubenswrapper[4872]: E0126 09:45:00.178343 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" containerName="registry-server" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.178355 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" containerName="registry-server" Jan 26 09:45:00 crc kubenswrapper[4872]: E0126 09:45:00.178392 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" containerName="extract-utilities" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.178407 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" containerName="extract-utilities" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.178745 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="8514cdec-1aa1-4da5-81b9-d4a58dda8bb2" containerName="registry-server" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.179891 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.185294 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.185318 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.196219 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh"] Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.276788 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpzdw\" (UniqueName: \"kubernetes.io/projected/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-kube-api-access-kpzdw\") pod \"collect-profiles-29490345-nkkgh\" (UID: \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.276875 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-secret-volume\") pod \"collect-profiles-29490345-nkkgh\" (UID: \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.277349 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-config-volume\") pod \"collect-profiles-29490345-nkkgh\" (UID: \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.379973 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-config-volume\") pod \"collect-profiles-29490345-nkkgh\" (UID: \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.380195 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-secret-volume\") pod \"collect-profiles-29490345-nkkgh\" (UID: \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.380232 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpzdw\" (UniqueName: \"kubernetes.io/projected/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-kube-api-access-kpzdw\") pod \"collect-profiles-29490345-nkkgh\" (UID: \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.381644 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-config-volume\") pod \"collect-profiles-29490345-nkkgh\" (UID: \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.397483 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-secret-volume\") pod \"collect-profiles-29490345-nkkgh\" (UID: \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.400838 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpzdw\" (UniqueName: \"kubernetes.io/projected/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-kube-api-access-kpzdw\") pod \"collect-profiles-29490345-nkkgh\" (UID: \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.507274 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:00 crc kubenswrapper[4872]: I0126 09:45:00.965675 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh"] Jan 26 09:45:01 crc kubenswrapper[4872]: I0126 09:45:01.598818 4872 generic.go:334] "Generic (PLEG): container finished" podID="8eb4e6c1-dd62-4494-b3e5-cfb1f6360477" containerID="e1bd60c3cffd895bfca7d1ed50e7ed288a90948d156687d518d0ccf3331d2144" exitCode=0 Jan 26 09:45:01 crc kubenswrapper[4872]: I0126 09:45:01.599117 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" event={"ID":"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477","Type":"ContainerDied","Data":"e1bd60c3cffd895bfca7d1ed50e7ed288a90948d156687d518d0ccf3331d2144"} Jan 26 09:45:01 crc kubenswrapper[4872]: I0126 09:45:01.599149 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" event={"ID":"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477","Type":"ContainerStarted","Data":"d077809b8c2f6200efee159fa87a9b1e094aab18430d863872e3e9a935dc3db3"} Jan 26 09:45:02 crc kubenswrapper[4872]: I0126 09:45:02.944340 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:03 crc kubenswrapper[4872]: I0126 09:45:03.035434 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpzdw\" (UniqueName: \"kubernetes.io/projected/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-kube-api-access-kpzdw\") pod \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\" (UID: \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\") " Jan 26 09:45:03 crc kubenswrapper[4872]: I0126 09:45:03.035512 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-config-volume\") pod \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\" (UID: \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\") " Jan 26 09:45:03 crc kubenswrapper[4872]: I0126 09:45:03.035611 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-secret-volume\") pod \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\" (UID: \"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477\") " Jan 26 09:45:03 crc kubenswrapper[4872]: I0126 09:45:03.036587 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-config-volume" (OuterVolumeSpecName: "config-volume") pod "8eb4e6c1-dd62-4494-b3e5-cfb1f6360477" (UID: "8eb4e6c1-dd62-4494-b3e5-cfb1f6360477"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 26 09:45:03 crc kubenswrapper[4872]: I0126 09:45:03.041241 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8eb4e6c1-dd62-4494-b3e5-cfb1f6360477" (UID: "8eb4e6c1-dd62-4494-b3e5-cfb1f6360477"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 26 09:45:03 crc kubenswrapper[4872]: I0126 09:45:03.041363 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-kube-api-access-kpzdw" (OuterVolumeSpecName: "kube-api-access-kpzdw") pod "8eb4e6c1-dd62-4494-b3e5-cfb1f6360477" (UID: "8eb4e6c1-dd62-4494-b3e5-cfb1f6360477"). InnerVolumeSpecName "kube-api-access-kpzdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 26 09:45:03 crc kubenswrapper[4872]: I0126 09:45:03.138444 4872 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-config-volume\") on node \"crc\" DevicePath \"\"" Jan 26 09:45:03 crc kubenswrapper[4872]: I0126 09:45:03.138478 4872 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-secret-volume\") on node \"crc\" DevicePath \"\"" Jan 26 09:45:03 crc kubenswrapper[4872]: I0126 09:45:03.138491 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpzdw\" (UniqueName: \"kubernetes.io/projected/8eb4e6c1-dd62-4494-b3e5-cfb1f6360477-kube-api-access-kpzdw\") on node \"crc\" DevicePath \"\"" Jan 26 09:45:03 crc kubenswrapper[4872]: I0126 09:45:03.615974 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" event={"ID":"8eb4e6c1-dd62-4494-b3e5-cfb1f6360477","Type":"ContainerDied","Data":"d077809b8c2f6200efee159fa87a9b1e094aab18430d863872e3e9a935dc3db3"} Jan 26 09:45:03 crc kubenswrapper[4872]: I0126 09:45:03.616015 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d077809b8c2f6200efee159fa87a9b1e094aab18430d863872e3e9a935dc3db3" Jan 26 09:45:03 crc kubenswrapper[4872]: I0126 09:45:03.616056 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29490345-nkkgh" Jan 26 09:45:04 crc kubenswrapper[4872]: I0126 09:45:04.052991 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs"] Jan 26 09:45:04 crc kubenswrapper[4872]: I0126 09:45:04.064354 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29490300-2v9qs"] Jan 26 09:45:05 crc kubenswrapper[4872]: I0126 09:45:05.204603 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2c67438-c0c7-48b8-a3df-21669ccb1602" path="/var/lib/kubelet/pods/f2c67438-c0c7-48b8-a3df-21669ccb1602/volumes" Jan 26 09:45:27 crc kubenswrapper[4872]: I0126 09:45:27.971081 4872 scope.go:117] "RemoveContainer" containerID="76664482753c3c2630369c1de9f8fe161bea6d4557a168d0d81b98f6c2e95d4c" Jan 26 09:45:40 crc kubenswrapper[4872]: I0126 09:45:40.228227 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:45:40 crc kubenswrapper[4872]: I0126 09:45:40.229017 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:46:10 crc kubenswrapper[4872]: I0126 09:46:10.227993 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:46:10 crc kubenswrapper[4872]: I0126 09:46:10.229174 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:46:40 crc kubenswrapper[4872]: I0126 09:46:40.227551 4872 patch_prober.go:28] interesting pod/machine-config-daemon-gt4gn container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Jan 26 09:46:40 crc kubenswrapper[4872]: I0126 09:46:40.228184 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Jan 26 09:46:40 crc kubenswrapper[4872]: I0126 09:46:40.228243 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" Jan 26 09:46:40 crc kubenswrapper[4872]: I0126 09:46:40.229056 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"56255d0196d693b2765251d758eac96be427715e4d4c8fd546a448498cf230c2"} pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Jan 26 09:46:40 crc kubenswrapper[4872]: I0126 09:46:40.229138 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" containerName="machine-config-daemon" containerID="cri-o://56255d0196d693b2765251d758eac96be427715e4d4c8fd546a448498cf230c2" gracePeriod=600 Jan 26 09:46:40 crc kubenswrapper[4872]: E0126 09:46:40.358851 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:46:40 crc kubenswrapper[4872]: I0126 09:46:40.630229 4872 generic.go:334] "Generic (PLEG): container finished" podID="fca44d96-a000-4bf2-8283-a937b0192880" containerID="56255d0196d693b2765251d758eac96be427715e4d4c8fd546a448498cf230c2" exitCode=0 Jan 26 09:46:40 crc kubenswrapper[4872]: I0126 09:46:40.630282 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" event={"ID":"fca44d96-a000-4bf2-8283-a937b0192880","Type":"ContainerDied","Data":"56255d0196d693b2765251d758eac96be427715e4d4c8fd546a448498cf230c2"} Jan 26 09:46:40 crc kubenswrapper[4872]: I0126 09:46:40.630322 4872 scope.go:117] "RemoveContainer" containerID="8dad67da6bdcbaf03f302de0fdcef76cd940f73a0eab7a62b9aac33c1f3c41aa" Jan 26 09:46:40 crc kubenswrapper[4872]: I0126 09:46:40.631470 4872 scope.go:117] "RemoveContainer" containerID="56255d0196d693b2765251d758eac96be427715e4d4c8fd546a448498cf230c2" Jan 26 09:46:40 crc kubenswrapper[4872]: E0126 09:46:40.631966 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:46:53 crc kubenswrapper[4872]: I0126 09:46:53.186838 4872 scope.go:117] "RemoveContainer" containerID="56255d0196d693b2765251d758eac96be427715e4d4c8fd546a448498cf230c2" Jan 26 09:46:53 crc kubenswrapper[4872]: E0126 09:46:53.187973 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" Jan 26 09:47:05 crc kubenswrapper[4872]: I0126 09:47:05.184945 4872 scope.go:117] "RemoveContainer" containerID="56255d0196d693b2765251d758eac96be427715e4d4c8fd546a448498cf230c2" Jan 26 09:47:05 crc kubenswrapper[4872]: E0126 09:47:05.186245 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-gt4gn_openshift-machine-config-operator(fca44d96-a000-4bf2-8283-a937b0192880)\"" pod="openshift-machine-config-operator/machine-config-daemon-gt4gn" podUID="fca44d96-a000-4bf2-8283-a937b0192880" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515135634055024454 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015135634056017372 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015135627027016515 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015135627030015457 5ustar corecore